- abort(Throwable) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
-
- abort(WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
-
- abort() - Method in interface org.apache.spark.sql.connector.write.DataWriter
-
Aborts this writer if it is failed.
- abort(long, WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
-
- abortStagedChanges() - Method in interface org.apache.spark.sql.connector.catalog.StagedTable
-
Abort the changes that were staged, both in metadata and from temporary outputs of this
table's writers.
- abs(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the absolute value of a numeric value.
- abs(T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
-
- abs() - Method in class org.apache.spark.sql.types.Decimal
-
- abs(T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
-
- abs(double) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
-
- abs(double) - Method in interface org.apache.spark.sql.types.DoubleType.DoubleIsConflicted
-
- abs(float) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
-
- abs(float) - Method in interface org.apache.spark.sql.types.FloatType.FloatIsConflicted
-
- abs(T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
-
- abs(T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
-
- abs(T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
-
- absent() - Static method in class org.apache.spark.api.java.Optional
-
- AbsoluteError - Class in org.apache.spark.mllib.tree.loss
-
Class for absolute error loss calculation (for regression).
- AbsoluteError() - Constructor for class org.apache.spark.mllib.tree.loss.AbsoluteError
-
- AbstractLauncher<T extends AbstractLauncher<T>> - Class in org.apache.spark.launcher
-
Base class for launcher implementations.
- accept(Parsers) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- accept(ES, Function1<ES, List<Object>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- accept(String, PartialFunction<Object, U>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- accept(Path) - Method in class org.apache.spark.ml.image.SamplePathFilter
-
- acceptIf(Function1<Object, Object>, Function1<Object, String>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- acceptMatch(String, PartialFunction<Object, U>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- acceptSeq(ES, Function1<ES, Iterable<Object>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- AcceptsLatestSeenOffset - Interface in org.apache.spark.sql.connector.read.streaming
-
Indicates that the source accepts the latest seen offset, which requires streaming execution
to provide the latest seen offset when restarting the streaming query from checkpoint.
- acceptsType(DataType) - Method in class org.apache.spark.sql.types.ObjectType
-
- accessNonExistentAccumulatorError(long) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- accId() - Method in class org.apache.spark.CleanAccum
-
- accumCleaned(long) - Method in interface org.apache.spark.CleanerListener
-
- AccumulableInfo - Class in org.apache.spark.scheduler
-
:: DeveloperApi ::
Information about an
AccumulatorV2
modified during a task or stage.
- AccumulableInfo - Class in org.apache.spark.status.api.v1
-
- accumulableInfoFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- AccumulableInfoSerializer - Class in org.apache.spark.status.protobuf
-
- AccumulableInfoSerializer() - Constructor for class org.apache.spark.status.protobuf.AccumulableInfoSerializer
-
- accumulableInfoToJson(AccumulableInfo, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- accumulables() - Method in class org.apache.spark.scheduler.StageInfo
-
Terminal values of accumulables updated during this stage, including all the user-defined
accumulators.
- accumulables() - Method in class org.apache.spark.scheduler.TaskInfo
-
Intermediate updates to accumulables during this task.
- accumulablesToJson(Iterable<AccumulableInfo>, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- ACCUMULATOR_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- ACCUMULATOR_UPDATES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- AccumulatorContext - Class in org.apache.spark.util
-
An internal class used to track accumulators by Spark itself.
- AccumulatorContext() - Constructor for class org.apache.spark.util.AccumulatorContext
-
- ACCUMULATORS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- accumulatorUpdates() - Method in class org.apache.spark.status.api.v1.StageData
-
- accumulatorUpdates() - Method in class org.apache.spark.status.api.v1.TaskData
-
- AccumulatorV2<IN,OUT> - Class in org.apache.spark.util
-
The base class for accumulators, that can accumulate inputs of type IN
, and produce output of
type OUT
.
- AccumulatorV2() - Constructor for class org.apache.spark.util.AccumulatorV2
-
- accumUpdates() - Method in class org.apache.spark.ExceptionFailure
-
- accumUpdates() - Method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- accumUpdates() - Method in class org.apache.spark.TaskKilled
-
- accuracy() - Method in interface org.apache.spark.ml.classification.ClassificationSummary
-
Returns accuracy.
- accuracy() - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
-
- accuracy() - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns accuracy
- acos(Column) - Static method in class org.apache.spark.sql.functions
-
- acos(String) - Static method in class org.apache.spark.sql.functions
-
- acosh(Column) - Static method in class org.apache.spark.sql.functions
-
- acosh(String) - Static method in class org.apache.spark.sql.functions
-
- acquire(Seq<String>) - Method in interface org.apache.spark.resource.ResourceAllocator
-
Acquire a sequence of resource addresses (to a launched task), these addresses must be
available.
- actionNotAllowedOnTableSincePartitionMetadataNotStoredError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- actionNotAllowedOnTableWithFilesourcePartitionManagementDisabledError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ActivationFunction - Interface in org.apache.spark.ml.ann
-
Trait for functions and their derivatives for functional layers
- active() - Static method in class org.apache.spark.sql.SparkSession
-
Returns the currently active SparkSession, otherwise the default one.
- active() - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns a list of active queries associated with this SQLContext
- active() - Method in class org.apache.spark.streaming.scheduler.ReceiverInfo
-
- ACTIVE() - Static method in class org.apache.spark.streaming.scheduler.ReceiverState
-
- ACTIVE_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- activeIterator() - Method in interface org.apache.spark.ml.linalg.Vector
-
Returns an iterator over all the active elements of this vector.
- activeIterator() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Returns an iterator over all the active elements of this vector.
- activeStages() - Method in class org.apache.spark.status.LiveJob
-
- activeTasks() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
-
- activeTasks() - Method in class org.apache.spark.status.LiveJob
-
- activeTasks() - Method in class org.apache.spark.status.LiveStage
-
- activeTasksPerExecutor() - Method in class org.apache.spark.status.LiveStage
-
- add(Tuple2<Vector, Object>) - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
-
Add a new training instance to this ExpectationAggregator, update the weights,
means and covariances for each distributions, and update the log likelihood.
- add(org.apache.spark.ml.feature.InstanceBlock) - Method in class org.apache.spark.ml.clustering.KMeansAggregator
-
- add(Term) - Static method in class org.apache.spark.ml.feature.Dot
-
- add(Term) - Static method in class org.apache.spark.ml.feature.EmptyTerm
-
- add(Term) - Method in interface org.apache.spark.ml.feature.Term
-
Creates a summation term by concatenation of terms.
- add(Datum) - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
-
Add a single data point to this aggregator.
- add(double[], MultivariateGaussian[], ExpectationSum, Vector<Object>) - Static method in class org.apache.spark.mllib.clustering.ExpectationSum
-
- add(Vector) - Method in class org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
Adds a new document.
- add(BlockMatrix) - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Adds the given block matrix other
to this
block matrix: this + other
.
- add(Vector) - Method in class org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Add a new sample to this summarizer, and update the statistical summary.
- add(StructField) - Method in class org.apache.spark.sql.types.StructType
-
- add(String, DataType) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new nullable field with no metadata.
- add(String, DataType, boolean) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field with no metadata.
- add(String, DataType, boolean, Metadata) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field and specifying metadata.
- add(String, DataType, boolean, String) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field and specifying metadata.
- add(String, String) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new nullable field with no metadata where the
dataType is specified as a String.
- add(String, String, boolean) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field with no metadata where the
dataType is specified as a String.
- add(String, String, boolean, Metadata) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field and specifying metadata where the
dataType is specified as a String.
- add(String, String, boolean, String) - Method in class org.apache.spark.sql.types.StructType
-
Creates a new
StructType
by adding a new field and specifying metadata where the
dataType is specified as a String.
- add(Long) - Method in class org.apache.spark.sql.util.MapperRowCounter
-
- add(double) - Method in class org.apache.spark.sql.util.NumericHistogram
-
Adds a new data point to the histogram approximation.
- add(T) - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- add(long, long) - Static method in class org.apache.spark.streaming.util.RawTextHelper
-
- add(IN) - Method in class org.apache.spark.util.AccumulatorV2
-
Takes the inputs and accumulates.
- add(T) - Method in class org.apache.spark.util.CollectionAccumulator
-
- add(Double) - Method in class org.apache.spark.util.DoubleAccumulator
-
Adds v to the accumulator, i.e.
- add(double) - Method in class org.apache.spark.util.DoubleAccumulator
-
Adds v to the accumulator, i.e.
- add(Long) - Method in class org.apache.spark.util.LongAccumulator
-
Adds v to the accumulator, i.e.
- add(long) - Method in class org.apache.spark.util.LongAccumulator
-
Adds v to the accumulator, i.e.
- add(Object) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by one.
- add(Object, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by count
.
- add_months(Column, int) - Static method in class org.apache.spark.sql.functions
-
Returns the date that is numMonths
after startDate
.
- add_months(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns the date that is numMonths
after startDate
.
- ADD_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- ADD_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdatesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAddresses(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- addAddressesBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAllAccumulatorUpdates(Iterable<? extends StoreTypes.AccumulableInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- addAllAddresses(Iterable<String>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- addAllAttempts(Iterable<? extends StoreTypes.ApplicationAttemptInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAllBlacklistedInStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- addAllBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- addAllBytesWritten(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- addAllChildClusters(Iterable<? extends StoreTypes.RDDOperationClusterWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addAllChildNodes(Iterable<? extends StoreTypes.RDDOperationNode>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addAllClasspathEntries(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addAllCorruptMergedBlockChunks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- addAllDataDistribution(Iterable<? extends StoreTypes.RDDDataDistribution>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addAllDiskBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- addAllDiskBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- addAllDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- addAllEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addAllEdges(Iterable<? extends StoreTypes.SparkPlanGraphEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addAllExcludedInStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- addAllExecutorCpuTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- addAllExecutorDeserializeCpuTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- addAllExecutorDeserializeTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- addAllExecutorMetrics(Iterable<? extends StoreTypes.ExecutorMetrics>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addAllExecutorRunTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- addAllExecutors(Iterable<String>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- addAllFailedTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- addAllFetchWaitTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- addAllGettingResultTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- addAllHadoopProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addAllIncomingEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addAllInputBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- addAllInputRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- addAllJobIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- addAllJobTags(Iterable<String>) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- addAllJvmGcTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- addAllKilledTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- addAllLocalBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- addAllLocalMergedBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- addAllLocalMergedBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- addAllLocalMergedChunksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- addAllMemoryBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- addAllMemoryBytesSpilled(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- addAllMergedFetchFallbackCount(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addAllMetrics(Iterable<? extends StoreTypes.SQLPlanMetric>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addAllMetricsProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addAllNodes(Iterable<? extends StoreTypes.SparkPlanGraphNodeWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addAllNodes(Iterable<? extends StoreTypes.SparkPlanGraphNodeWrapper>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addAllOutgoingEdges(Iterable<? extends StoreTypes.RDDOperationEdge>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addAllOutputBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- addAllOutputRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- addAllPartitions(Iterable<? extends StoreTypes.RDDPartitionInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addAllPeakExecutionMemory(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- addAllQuantiles(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- addAllRddIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- addAllReadBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- addAllReadRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- addAllRecordsRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- addAllRecordsWritten(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- addAllRemoteBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- addAllRemoteBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- addAllRemoteBytesReadToDisk(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- addAllRemoteMergedBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- addAllRemoteMergedBytesRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- addAllRemoteMergedChunksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- addAllRemoteMergedReqsDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- addAllRemoteReqsDuration(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- addAllResourceProfiles(Iterable<? extends StoreTypes.ResourceProfileInfo>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addAllResultSerializationTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- addAllResultSize(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- addAllSchedulerDelay(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- addAllShuffleRead(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- addAllShuffleReadRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- addAllShuffleWrite(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- addAllShuffleWriteRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- addAllSkippedStages(Iterable<? extends Integer>) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- addAllSources(Iterable<? extends StoreTypes.SourceProgress>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addAllSparkProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addAllStageIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- addAllStageIds(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- addAllStages(Iterable<? extends Long>) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- addAllStateOperators(Iterable<? extends StoreTypes.StateOperatorProgress>) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addAllSucceededTasks(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- addAllSystemProperties(Iterable<? extends StoreTypes.PairStrings>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addAllTaskTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- addAllTotalBlocksFetched(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- addAllWriteBytes(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- addAllWriteRecords(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- addAllWriteTime(Iterable<? extends Double>) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- addAppArgs(String...) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds command line arguments for the application.
- addAppArgs(String...) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addArchive(String) - Method in class org.apache.spark.SparkContext
-
:: Experimental ::
Add an archive to be downloaded and unpacked with this Spark job on every node.
- addAttempts(StoreTypes.ApplicationAttemptInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAttempts(int, StoreTypes.ApplicationAttemptInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAttempts(StoreTypes.ApplicationAttemptInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAttempts(int, StoreTypes.ApplicationAttemptInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAttemptsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addAttemptsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- addBin(double, double, int) - Method in class org.apache.spark.sql.util.NumericHistogram
-
Set a particular histogram bin with index.
- addBinary(byte[]) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by one.
- addBinary(byte[], long) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by count
.
- addBlacklistedInStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- addBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- addBytesWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- addCatalogInCacheTableAsSelectNotAllowedError(String, SqlBaseParser.CacheTableContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- addChildClusters(StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildClusters(int, StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildClusters(StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildClusters(int, StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildClustersBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildClustersBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- addChildNodes(StoreTypes.RDDOperationNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChildNodes(int, StoreTypes.RDDOperationNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChildNodes(StoreTypes.RDDOperationNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChildNodes(int, StoreTypes.RDDOperationNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChildNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChildNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- addChunk(ShuffleBlockChunkId, RoaringBitmap) - Method in class org.apache.spark.storage.PushBasedFetchHelper
-
This is executed by the task thread when the iterator.next()
is invoked and the iterator
processes a response of type ShuffleBlockFetcherIterator.PushMergedLocalMetaFetchResult
.
- addClasspathEntries(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addClasspathEntries(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addClasspathEntries(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addClasspathEntries(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addClasspathEntriesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addClasspathEntriesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- addColumn(String[], DataType) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
-
Create a TableChange for adding an optional column.
- addColumn(String[], DataType, boolean) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
-
Create a TableChange for adding a column.
- addColumn(String[], DataType, boolean, String) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
-
Create a TableChange for adding a column.
- addColumn(String[], DataType, boolean, String, TableChange.ColumnPosition, ColumnDefaultValue) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
-
Create a TableChange for adding a column.
- addColumnWithV1TableCannotSpecifyNotNullError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- addCorruptMergedBlockChunks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- addDataDistribution(StoreTypes.RDDDataDistribution) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDataDistribution(int, StoreTypes.RDDDataDistribution) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDataDistribution(StoreTypes.RDDDataDistribution.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDataDistribution(int, StoreTypes.RDDDataDistribution.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDataDistributionBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDataDistributionBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- addDirectoryError(Path) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- addDiskBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- addDiskBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- addDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- addEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdges(StoreTypes.SparkPlanGraphEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addEdges(int, StoreTypes.SparkPlanGraphEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addEdges(StoreTypes.SparkPlanGraphEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addEdges(int, StoreTypes.SparkPlanGraphEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- addEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- addExcludedInStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- addExecutorCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- addExecutorDeserializeCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- addExecutorDeserializeTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- addExecutorMetrics(StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorMetrics(int, StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorMetrics(StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorMetrics(int, StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- addExecutorRunTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- addExecutors(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- addExecutorsBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- addFailedTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- addFetchWaitTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- addFile(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String, boolean) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds a file to be submitted with the application.
- addFile(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addFile(String) - Method in class org.apache.spark.SparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String, boolean) - Method in class org.apache.spark.SparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFilesWithAbsolutePathUnsupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- addFilter(ServletContextHandler, String, Map<String, String>) - Static method in class org.apache.spark.ui.JettyUtils
-
- addGettingResultTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- addGrid(Param<T>, Iterable<T>) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a param with multiple values (overwrites if the input param exists).
- addGrid(DoubleParam, double[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a double param with multiple values.
- addGrid(IntParam, int[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds an int param with multiple values.
- addGrid(FloatParam, float[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a float param with multiple values.
- addGrid(LongParam, long[]) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a long param with multiple values.
- addGrid(BooleanParam) - Method in class org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a boolean param with true and false.
- addHadoopProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addHadoopProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addHadoopProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addHadoopProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addHadoopPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addHadoopPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- addIncomingEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addIncomingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addIncomingEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addIncomingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addIncomingEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addIncomingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- addInputBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- addInputRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- addJar(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Adds a JAR dependency for all tasks to be executed on this SparkContext in the future.
- addJar(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds a jar file to be submitted with the application.
- addJar(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addJar(String) - Method in class org.apache.spark.SparkContext
-
Adds a JAR dependency for all tasks to be executed on this SparkContext
in the future.
- addJarsToClassPath(String, MutableURLClassLoader) - Static method in class org.apache.spark.util.DependencyUtils
-
- addJarToClasspath(String, MutableURLClassLoader) - Static method in class org.apache.spark.util.DependencyUtils
-
- addJobIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- addJobTag(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Add a tag to be assigned to all the jobs started by this thread.
- addJobTag(String) - Method in class org.apache.spark.SparkContext
-
Add a tag to be assigned to all the jobs started by this thread.
- addJobTags(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- addJobTagsBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- addJvmGcTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- addKilledTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- addListener(SparkAppHandle.Listener) - Method in interface org.apache.spark.launcher.SparkAppHandle
-
Adds a listener to be notified of changes to the handle's information.
- addListener(StreamingQueryListener) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
- addListener(L) - Method in interface org.apache.spark.util.ListenerBus
-
Add a listener to listen events.
- addLocalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- addLocalConfiguration(String, int, int, int, JobConf) - Static method in class org.apache.spark.rdd.HadoopRDD
-
Add Hadoop configuration specific to a single partition and attempt.
- addLocalDirectoryError(Path) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- addLocalMergedBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- addLocalMergedBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- addLocalMergedChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- addLong(long) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by one.
- addLong(long, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by count
.
- addMapOutput(int, MapStatus) - Method in class org.apache.spark.ShuffleStatus
-
Register a map output.
- addMemoryBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- addMemoryBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- addMergedFetchFallbackCount(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- addMergeResult(int, org.apache.spark.scheduler.MergeStatus) - Method in class org.apache.spark.ShuffleStatus
-
Register a merge result.
- addMetrics(TaskMetrics, TaskMetrics) - Static method in class org.apache.spark.status.LiveEntityHelpers
-
Add m2 values to m1.
- addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetrics(StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetrics(StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- addMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- addMetricsProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addMetricsProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addMetricsProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addMetricsProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addMetricsPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addMetricsPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- addNaN() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- addNewDefaultColumnToExistingTableNotAllowed(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- addNewFunctionMismatchedWithFunctionError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- addNodes(StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodes(StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodes(StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNodes(StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- addNodesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- addNull() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- addOutgoingEdges(StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutgoingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutgoingEdges(StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutgoingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutgoingEdgesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutgoingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- addOutputBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- addOutputRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- addPartition(LiveRDDPartition) - Method in class org.apache.spark.status.RDDPartitionSeq
-
- addPartitions(StoreTypes.RDDPartitionInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartitions(int, StoreTypes.RDDPartitionInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartitions(StoreTypes.RDDPartitionInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartitions(int, StoreTypes.RDDPartitionInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartitionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartitionsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- addPartToPGroup(Partition, PartitionGroup) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
- addPeakExecutionMemory(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- addPyFile(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds a python file / zip / egg to be submitted with the application.
- addPyFile(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- addQuantiles(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- addRddIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- addReadBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- addReadRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- addRecordsRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- addRecordsWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- addRemoteBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- addRemoteBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- addRemoteBytesReadToDisk(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- addRemoteMergedBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- addRemoteMergedBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- addRemoteMergedChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- addRemoteMergedReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- addRemoteReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- addRepeatedField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- addRequest(TaskResourceRequest) - Method in class org.apache.spark.resource.TaskResourceRequests
-
- addResourceProfiles(StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addResourceProfiles(int, StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addResourceProfiles(StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addResourceProfiles(int, StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addResourceProfilesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- addResourceProfilesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- address() - Method in class org.apache.spark.BarrierTaskInfo
-
- address() - Method in class org.apache.spark.status.api.v1.RDDDataDistribution
-
- ADDRESS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- addresses() - Method in class org.apache.spark.resource.ResourceInformation
-
- addresses() - Method in class org.apache.spark.resource.ResourceInformationJson
-
- ADDRESSES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- addResultSerializationTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- addResultSize(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- addSchedulable(Schedulable) - Method in interface org.apache.spark.scheduler.Schedulable
-
- addSchedulerDelay(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- addShuffleRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- addShuffleReadRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- addShuffleWrite(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- addShuffleWriteRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- addShutdownHook(Function0<BoxedUnit>) - Static method in class org.apache.spark.util.ShutdownHookManager
-
Adds a shutdown hook with default priority.
- addShutdownHook(int, Function0<BoxedUnit>) - Static method in class org.apache.spark.util.ShutdownHookManager
-
Adds a shutdown hook with the given priority.
- addSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- addSources(StoreTypes.SourceProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSources(int, StoreTypes.SourceProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSources(StoreTypes.SourceProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSources(int, StoreTypes.SourceProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSourcesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSourcesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- addSparkArg(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds a no-value argument to the Spark invocation.
- addSparkArg(String, String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Adds an argument with a value to the Spark invocation.
- addSparkArg(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addSparkArg(String, String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- addSparkListener(SparkListenerInterface) - Method in class org.apache.spark.SparkContext
-
:: DeveloperApi ::
Register a listener to receive up-calls from events that happen during execution.
- addSparkProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addSparkProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addSparkProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addSparkProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addSparkPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addSparkPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- addStageIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- addStageIds(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- addStages(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- addStateOperators(StoreTypes.StateOperatorProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStateOperators(int, StoreTypes.StateOperatorProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStateOperators(StoreTypes.StateOperatorProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStateOperators(int, StoreTypes.StateOperatorProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStateOperatorsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStateOperatorsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- addStreamingListener(StreamingListener) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
- addStreamingListener(StreamingListener) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
- addString(String) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by one.
- addString(String, long) - Method in class org.apache.spark.util.sketch.CountMinSketch
-
Increments item
's count by count
.
- addSucceededTasks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- addSystemProperties(StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addSystemProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addSystemProperties(StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addSystemProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addSystemPropertiesBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addSystemPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- addTaskCompletionListener(TaskCompletionListener) - Method in class org.apache.spark.BarrierTaskContext
-
- addTaskCompletionListener(TaskCompletionListener) - Method in class org.apache.spark.TaskContext
-
Adds a (Java friendly) listener to be executed on task completion.
- addTaskCompletionListener(Function1<TaskContext, U>) - Method in class org.apache.spark.TaskContext
-
Adds a listener in the form of a Scala closure to be executed on task completion.
- addTaskFailureListener(TaskFailureListener) - Method in class org.apache.spark.BarrierTaskContext
-
- addTaskFailureListener(TaskFailureListener) - Method in class org.apache.spark.TaskContext
-
Adds a listener to be executed on task failure (which includes completion listener failure, if
the task body did not already fail).
- addTaskFailureListener(Function2<TaskContext, Throwable, BoxedUnit>) - Method in class org.apache.spark.TaskContext
-
Adds a listener to be executed on task failure (which includes completion listener failure, if
the task body did not already fail).
- addTaskResourceRequests(SparkConf, TaskResourceRequests) - Static method in class org.apache.spark.resource.ResourceUtils
-
- addTaskSetManager(Schedulable, Properties) - Method in interface org.apache.spark.scheduler.SchedulableBuilder
-
- addTaskTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- addTime() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
-
- addTime() - Method in class org.apache.spark.status.api.v1.ProcessSummary
-
- addTotalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- addURL(URL) - Method in class org.apache.spark.util.MutableURLClassLoader
-
- AddWebUIFilter(String, Map<String, String>, String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
-
- AddWebUIFilter$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter$
-
- addWriteBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- addWriteRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- addWriteTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- advisoryPartitionSizeInBytes() - Method in interface org.apache.spark.sql.connector.write.RequiresDistributionAndOrdering
-
Returns the advisory (not guaranteed) shuffle partition size in bytes for this write.
- aes_decrypt(Column, Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a decrypted value of input
using AES in mode
with padding
.
- aes_decrypt(Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a decrypted value of input
.
- aes_decrypt(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a decrypted value of input
.
- aes_decrypt(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a decrypted value of input
.
- aes_encrypt(Column, Column, Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an encrypted value of input
using AES in given mode
with the specified padding
.
- aes_encrypt(Column, Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an encrypted value of input
.
- aes_encrypt(Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an encrypted value of input
.
- aes_encrypt(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an encrypted value of input
.
- aes_encrypt(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an encrypted value of input
.
- aesCryptoError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- aesModeUnsupportedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- aesUnsupportedAad(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- aesUnsupportedIv(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- after(String) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange.ColumnPosition
-
- AFTSurvivalRegression - Class in org.apache.spark.ml.regression
-
- AFTSurvivalRegression(String) - Constructor for class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- AFTSurvivalRegression() - Constructor for class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- AFTSurvivalRegressionModel - Class in org.apache.spark.ml.regression
-
- AFTSurvivalRegressionParams - Interface in org.apache.spark.ml.regression
-
Params for accelerated failure time (AFT) regression.
- agg(Column, Column...) - Method in class org.apache.spark.sql.Dataset
-
Aggregates on the entire Dataset without groups.
- agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - Method in class org.apache.spark.sql.Dataset
-
(Scala-specific) Aggregates on the entire Dataset without groups.
- agg(Map<String, String>) - Method in class org.apache.spark.sql.Dataset
-
(Scala-specific) Aggregates on the entire Dataset without groups.
- agg(Map<String, String>) - Method in class org.apache.spark.sql.Dataset
-
(Java-specific) Aggregates on the entire Dataset without groups.
- agg(Column, Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Aggregates on the entire Dataset without groups.
- agg(TypedColumn<V, U1>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregation, returning a
Dataset
of tuples for each unique key
and the result of computing this aggregation over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>, TypedColumn<V, U7>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>, TypedColumn<V, U5>, TypedColumn<V, U6>, TypedColumn<V, U7>, TypedColumn<V, U8>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset
of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(Column, Column...) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute aggregates by specifying a series of aggregate columns.
- agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
(Scala-specific) Compute aggregates by specifying the column names and
aggregate methods.
- agg(Map<String, String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
(Scala-specific) Compute aggregates by specifying a map from column name to
aggregate methods.
- agg(Map<String, String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
(Java-specific) Compute aggregates by specifying a map from column name to
aggregate methods.
- agg(Column, Seq<Column>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute aggregates by specifying a series of aggregate columns.
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Aggregate the elements of each partition, and then the results for all the partitions, using
given combine functions and a neutral "zero value".
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
-
Aggregate the elements of each partition, and then the results for all the partitions, using
given combine functions and a neutral "zero value".
- aggregate(Column, Column, Function2<Column, Column, Column>, Function1<Column, Column>) - Static method in class org.apache.spark.sql.functions
-
Applies a binary operator to an initial state and all elements in the array,
and reduces this to a single state.
- aggregate(Column, Column, Function2<Column, Column, Column>) - Static method in class org.apache.spark.sql.functions
-
Applies a binary operator to an initial state and all elements in the array,
and reduces this to a single state.
- aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- AggregatedDialect - Class in org.apache.spark.sql.jdbc
-
AggregatedDialect can unify multiple dialects into one virtual Dialect.
- AggregatedDialect(List<JdbcDialect>) - Constructor for class org.apache.spark.sql.jdbc.AggregatedDialect
-
- aggregateExpressionRequiredForPivotError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- aggregateExpressions() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
-
- AggregateFunc - Interface in org.apache.spark.sql.connector.expressions.aggregate
-
Base class of the Aggregate Functions.
- AggregateFunction<S extends java.io.Serializable,R> - Interface in org.apache.spark.sql.connector.catalog.functions
-
Interface for a function that produces a result value by aggregating over multiple input rows.
- aggregateInAggregateFilterError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- aggregateMessages(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, ClassTag<A>) - Method in class org.apache.spark.graphx.Graph
-
Aggregates values from the neighboring edges and vertices of each vertex.
- aggregateMessagesWithActiveSet(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, Option<Tuple2<VertexRDD<?>, EdgeDirection>>, ClassTag<A>) - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- aggregateTaskMetrics(long[]) - Method in class org.apache.spark.sql.connector.metric.CustomAvgMetric
-
- aggregateTaskMetrics(long[]) - Method in interface org.apache.spark.sql.connector.metric.CustomMetric
-
Given an array of task metric values, returns aggregated final metric value.
- aggregateTaskMetrics(long[]) - Method in class org.apache.spark.sql.connector.metric.CustomSumMetric
-
- aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
- aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - Method in class org.apache.spark.graphx.VertexRDD
-
Aggregates vertices in messages
that have the same ids using reduceFunc
, returning a
VertexRDD co-indexed with this
.
- AggregatingEdgeContext<VD,ED,A> - Class in org.apache.spark.graphx.impl
-
- AggregatingEdgeContext(Function2<A, A, A>, Object, BitSet) - Constructor for class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- Aggregation - Class in org.apache.spark.sql.connector.expressions.aggregate
-
Aggregation in SQL statement.
- Aggregation(AggregateFunc[], Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
-
- aggregationDepth() - Method in class org.apache.spark.ml.classification.LinearSVC
-
- aggregationDepth() - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
- aggregationDepth() - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- aggregationDepth() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- aggregationDepth() - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- aggregationDepth() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- aggregationDepth() - Method in interface org.apache.spark.ml.param.shared.HasAggregationDepth
-
Param for suggested depth for treeAggregate (>= 2).
- aggregationDepth() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- aggregationDepth() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- aggregationDepth() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- aggregationDepth() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- aggregationDepth() - Method in class org.apache.spark.ml.regression.LinearRegression
-
- aggregationDepth() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- aggregationFunctionAppliedOnNonNumericColumnError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- aggregationFunctionAppliedOnNonNumericColumnError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- aggregationNotAllowedInMergeCondition(String, Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- Aggregator<K,V,C> - Class in org.apache.spark
-
:: DeveloperApi ::
A set of functions used to aggregate data.
- Aggregator(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Constructor for class org.apache.spark.Aggregator
-
- aggregator() - Method in class org.apache.spark.ShuffleDependency
-
- Aggregator<IN,BUF,OUT> - Class in org.apache.spark.sql.expressions
-
A base class for user-defined aggregations, which can be used in Dataset
operations to take
all of the elements of a group and reduce them to a single value.
- Aggregator() - Constructor for class org.apache.spark.sql.expressions.Aggregator
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
- aic() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
- Algo - Class in org.apache.spark.mllib.tree.configuration
-
Enum to select the algorithm for the decision tree
- Algo() - Constructor for class org.apache.spark.mllib.tree.configuration.Algo
-
- algo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- algo() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- algo() - Method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- algo() - Method in class org.apache.spark.mllib.tree.model.RandomForestModel
-
- algorithm() - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
- alias(String) - Method in class org.apache.spark.sql.Column
-
Gives the column an alias.
- alias(String) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with an alias set.
- alias(Symbol) - Method in class org.apache.spark.sql.Dataset
-
(Scala-specific) Returns a new Dataset with an alias set.
- aliasesNumberNotMatchUDTFOutputError(int, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- aliasNumberNotMatchColumnNumberError(int, int, TreeNode<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- All - Static variable in class org.apache.spark.graphx.TripletFields
-
Expose all the fields (source, edge, and destination).
- ALL_GATHER() - Static method in class org.apache.spark.RequestMethod
-
- ALL_REMOVALS_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- ALL_UPDATES_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- allAvailable() - Static method in interface org.apache.spark.sql.connector.read.streaming.ReadLimit
-
- allGather(String) - Method in class org.apache.spark.BarrierTaskContext
-
:: Experimental ::
Blocks until all tasks in the same stage have reached this routine.
- AllJobsCancelled - Class in org.apache.spark.scheduler
-
- AllJobsCancelled() - Constructor for class org.apache.spark.scheduler.AllJobsCancelled
-
- allocate(int) - Method in class org.apache.spark.sql.util.NumericHistogram
-
Sets the number of histogram bins to use for approximating data.
- allocator() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
-
- AllReceiverIds - Class in org.apache.spark.streaming.scheduler
-
A message used by ReceiverTracker to ask all receiver's ids still stored in
ReceiverTrackerEndpoint.
- AllReceiverIds() - Constructor for class org.apache.spark.streaming.scheduler.AllReceiverIds
-
- allRemovalsTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
-
- allSources() - Static method in class org.apache.spark.metrics.source.StaticSources
-
The set of all static sources.
- allSupportedExecutorResources() - Static method in class org.apache.spark.resource.ResourceProfile
-
Return all supported Spark built-in executor resources, custom resources like GPUs/FPGAs
are excluded.
- allUpdatesTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
-
- alpha() - Method in class org.apache.spark.ml.recommendation.ALS
-
- alpha() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
Param for the alpha parameter in the implicit preference formulation (nonnegative).
- alpha() - Method in class org.apache.spark.mllib.random.WeibullGenerator
-
- ALS - Class in org.apache.spark.ml.recommendation
-
Alternating Least Squares (ALS) matrix factorization.
- ALS(String) - Constructor for class org.apache.spark.ml.recommendation.ALS
-
- ALS() - Constructor for class org.apache.spark.ml.recommendation.ALS
-
- ALS - Class in org.apache.spark.mllib.recommendation
-
Alternating Least Squares matrix factorization.
- ALS() - Constructor for class org.apache.spark.mllib.recommendation.ALS
-
Constructs an ALS instance with default parameters: {numBlocks: -1, rank: 10, iterations: 10,
lambda: 0.01, implicitPrefs: false, alpha: 1.0}.
- ALS.InBlock$ - Class in org.apache.spark.ml.recommendation
-
- ALS.LeastSquaresNESolver - Interface in org.apache.spark.ml.recommendation
-
Trait for least squares solvers applied to the normal equation.
- ALS.Rating<ID> - Class in org.apache.spark.ml.recommendation
-
Rating class for better code readability.
- ALS.Rating$ - Class in org.apache.spark.ml.recommendation
-
- ALS.RatingBlock$ - Class in org.apache.spark.ml.recommendation
-
- ALSModel - Class in org.apache.spark.ml.recommendation
-
Model fitted by ALS.
- ALSModelParams - Interface in org.apache.spark.ml.recommendation
-
Common params for ALS and ALSModel.
- ALSParams - Interface in org.apache.spark.ml.recommendation
-
Common params for ALS.
- alterAddColNotSupportDatasourceTableError(Object, TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterAddColNotSupportViewError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterColumnCannotFindColumnInV1TableError(String, org.apache.spark.sql.connector.catalog.V1Table) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterColumnWithV1TableCannotSpecifyNotNullError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterDatabaseLocationUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterNamespace(String[], NamespaceChange...) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- alterNamespace(String[], NamespaceChange...) - Method in interface org.apache.spark.sql.connector.catalog.SupportsNamespaces
-
Apply a set of metadata changes to a namespace in the catalog.
- alterTable(Identifier, TableChange...) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- alterTable(Identifier, TableChange...) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
-
Apply a set of
changes
to a table in the catalog.
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- alterTable(String, Seq<TableChange>, int) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Alter an existing table.
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- alterTable(String, Seq<TableChange>, int) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- alterTableChangeColumnNotSupportedForColumnTypeError(String, StructField, StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterTableRecoverPartitionsNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterTableSerDePropertiesNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterTableSetSerdeForSpecificPartitionNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterTableSetSerdeNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterTableWithDropPartitionAndPurgeUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- alterV2TableSetLocationWithPartitionNotSupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- alterView(Identifier, ViewChange...) - Method in interface org.apache.spark.sql.connector.catalog.ViewCatalog
-
Apply
changes
to a view in the catalog.
- AlwaysFalse - Class in org.apache.spark.sql.connector.expressions.filter
-
A predicate that always evaluates to false
.
- AlwaysFalse() - Constructor for class org.apache.spark.sql.connector.expressions.filter.AlwaysFalse
-
- AlwaysFalse - Class in org.apache.spark.sql.sources
-
A filter that always evaluates to false
.
- AlwaysFalse() - Constructor for class org.apache.spark.sql.sources.AlwaysFalse
-
- AlwaysTrue - Class in org.apache.spark.sql.connector.expressions.filter
-
A predicate that always evaluates to true
.
- AlwaysTrue() - Constructor for class org.apache.spark.sql.connector.expressions.filter.AlwaysTrue
-
- AlwaysTrue - Class in org.apache.spark.sql.sources
-
A filter that always evaluates to true
.
- AlwaysTrue() - Constructor for class org.apache.spark.sql.sources.AlwaysTrue
-
- am() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager
-
- ambiguousAttributesInSelfJoinError(Seq<AttributeReference>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousColumnOrFieldError(Seq<String>, int) - Method in interface org.apache.spark.sql.errors.CompilationErrors
-
- ambiguousColumnOrFieldError(Seq<String>, int, Origin) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- ambiguousColumnOrFieldError(Seq<String>, int, Origin) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousLateralColumnAliasError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousLateralColumnAliasError(Seq<String>, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousReferenceError(String, Seq<Attribute>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousReferenceToFieldsError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ambiguousRelationAliasNameInNestedCTEError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- amount() - Method in class org.apache.spark.resource.ExecutorResourceRequest
-
- amount() - Method in class org.apache.spark.resource.ResourceRequest
-
- AMOUNT() - Static method in class org.apache.spark.resource.ResourceUtils
-
- amount() - Method in class org.apache.spark.resource.TaskResourceRequest
-
- AMOUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- AMOUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- AnalysisException - Exception in org.apache.spark.sql
-
Thrown when a query fails to analyze, usually because the query itself is invalid.
- AnalysisException(String, Map<String, String>, Option<Throwable>) - Constructor for exception org.apache.spark.sql.AnalysisException
-
- AnalysisException(String, Map<String, String>, QueryContext[], String) - Constructor for exception org.apache.spark.sql.AnalysisException
-
- AnalysisException(String, Map<String, String>) - Constructor for exception org.apache.spark.sql.AnalysisException
-
- AnalysisException(String, Map<String, String>, Origin) - Constructor for exception org.apache.spark.sql.AnalysisException
-
- AnalysisException(String, Map<String, String>, Origin, Option<Throwable>) - Constructor for exception org.apache.spark.sql.AnalysisException
-
- analyzeTableNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- analyzeTableNotSupportedOnViewsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- analyzingColumnStatisticsNotSupportedForColumnTypeError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- and(Column) - Method in class org.apache.spark.sql.Column
-
Boolean AND.
- And - Class in org.apache.spark.sql.connector.expressions.filter
-
A predicate that evaluates to true
iff both left
and right
evaluate to
true
.
- And(Predicate, Predicate) - Constructor for class org.apache.spark.sql.connector.expressions.filter.And
-
- And - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff both left
or right
evaluate to true
.
- And(Filter, Filter) - Constructor for class org.apache.spark.sql.sources.And
-
- ANOVATest - Class in org.apache.spark.ml.stat
-
ANOVA Test for continuous data.
- ANOVATest() - Constructor for class org.apache.spark.ml.stat.ANOVATest
-
- ansiDateTimeError(Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- ansiDateTimeParseError(Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- ansiIllegalArgumentError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- ansiIllegalArgumentError(IllegalArgumentException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- antecedent() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
-
- ANY() - Static method in class org.apache.spark.scheduler.TaskLocality
-
- any(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns true if at least one value of e
is true.
- any_value(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns some value of e
for a group of rows.
- any_value(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns some value of e
for a group of rows.
- AnyDataType - Class in org.apache.spark.sql.types
-
An AbstractDataType
that matches any concrete data types.
- AnyDataType() - Constructor for class org.apache.spark.sql.types.AnyDataType
-
- anyNull() - Method in interface org.apache.spark.sql.Row
-
Returns true if there are any NULL values in this row.
- anyNull() - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- anyNull() - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- AnyTimestampType - Class in org.apache.spark.sql.types
-
- AnyTimestampType() - Constructor for class org.apache.spark.sql.types.AnyTimestampType
-
- AnyTimestampTypeExpression - Class in org.apache.spark.sql.types
-
- AnyTimestampTypeExpression() - Constructor for class org.apache.spark.sql.types.AnyTimestampTypeExpression
-
- ApiHelper - Class in org.apache.spark.ui.jobs
-
- ApiHelper() - Constructor for class org.apache.spark.ui.jobs.ApiHelper
-
- ApiRequestContext - Interface in org.apache.spark.status.api.v1
-
- APP_SPARK_VERSION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- appAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- append() - Method in class org.apache.spark.sql.DataFrameWriterV2
-
Append the contents of the data frame to the output table.
- Append() - Static method in class org.apache.spark.sql.streaming.OutputMode
-
OutputMode in which only the new rows in the streaming DataFrame/Dataset will be
written to the sink.
- appendBias(Vector) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Returns a new vector with 1.0
(bias) appended to the input vector.
- appendColumn(StructType, String, DataType, boolean) - Static method in class org.apache.spark.ml.util.SchemaUtils
-
Appends a new column to the input schema.
- appendColumn(StructType, StructField) - Static method in class org.apache.spark.ml.util.SchemaUtils
-
Appends a new column to the input schema.
- AppHistoryServerPlugin - Interface in org.apache.spark.status
-
An interface for creating history listeners(to replay event logs) defined in other modules like
SQL, and setup the UI of the plugin to rebuild the history UI.
- appId() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- appId() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
-
- appId() - Method in class org.apache.spark.storage.ShuffleMergedDataBlockId
-
- appId() - Method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
-
- appId() - Method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
-
- APPLICATION_EXECUTOR_LIMIT() - Static method in class org.apache.spark.ui.ToolTips
-
- APPLICATION_MASTER() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
-
- applicationAttemptId() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
Get the attempt ID for this run, if the cluster manager supports multiple
attempts.
- applicationAttemptId() - Method in interface org.apache.spark.scheduler.TaskScheduler
-
Get an application's attempt ID associated with the job.
- applicationAttemptId() - Method in class org.apache.spark.SparkContext
-
- ApplicationAttemptInfo - Class in org.apache.spark.status.api.v1
-
- applicationEndFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- applicationEndToJson(SparkListenerApplicationEnd, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- ApplicationEnvironmentInfo - Class in org.apache.spark.status.api.v1
-
- applicationId() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
Get an application ID associated with the job.
- applicationId() - Method in interface org.apache.spark.scheduler.TaskScheduler
-
Get an application ID associated with the job.
- applicationId() - Method in class org.apache.spark.SparkContext
-
A unique identifier for the Spark application.
- ApplicationInfo - Class in org.apache.spark.status.api.v1
-
- APPLICATIONS() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
-
- applicationStartFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- applicationStartToJson(SparkListenerApplicationStart, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- ApplicationStatus - Enum in org.apache.spark.status.api.v1
-
- apply(T1) - Static method in class org.apache.spark.CleanAccum
-
- apply(T1) - Static method in class org.apache.spark.CleanBroadcast
-
- apply(T1) - Static method in class org.apache.spark.CleanCheckpoint
-
- apply(T1) - Static method in class org.apache.spark.CleanRDD
-
- apply(T1) - Static method in class org.apache.spark.CleanShuffle
-
- apply(T1) - Static method in class org.apache.spark.CleanSparkListener
-
- apply(T1, T2) - Static method in class org.apache.spark.ContextBarrierId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.ErrorInfo
-
- apply(int) - Static method in class org.apache.spark.ErrorMessageFormat
-
- apply(T1) - Static method in class org.apache.spark.ErrorSubInfo
-
- apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.ExceptionFailure
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.ExecutorLostFailure
-
- apply(T1) - Static method in class org.apache.spark.ExecutorRegistered
-
- apply(T1) - Static method in class org.apache.spark.ExecutorRemoved
-
- apply(T1, T2, T3, T4, T5, T6) - Static method in class org.apache.spark.FetchFailed
-
- apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.Graph
-
Construct a graph from a collection of vertices and
edges with attributes.
- apply(RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from edges, setting referenced vertices to defaultVertexAttr
.
- apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from vertices and edges, setting missing vertices to defaultVertexAttr
.
- apply(VertexRDD<VD>, EdgeRDD<ED>, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from a VertexRDD and an EdgeRDD with arbitrary replicated vertices.
- apply(Graph<VD, ED>, A, int, EdgeDirection, Function3<Object, VD, A, VD>, Function1<EdgeTriplet<VD, ED>, Iterator<Tuple2<Object, A>>>, Function2<A, A, A>, ClassTag<VD>, ClassTag<ED>, ClassTag<A>) - Static method in class org.apache.spark.graphx.Pregel
-
Execute a Pregel-like iterative vertex-parallel abstraction.
- apply(RDD<Tuple2<Object, VD>>, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
-
Constructs a standalone
VertexRDD
(one that is not set up for efficient joins with an
EdgeRDD
) from an RDD of vertex-attribute pairs.
- apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
-
Constructs a VertexRDD
from an RDD of vertex-attribute pairs.
- apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, Function2<VD, VD, VD>, ClassTag<VD>) - Static method in class org.apache.spark.graphx.VertexRDD
-
Constructs a VertexRDD
from an RDD of vertex-attribute pairs.
- apply(DenseMatrix<Object>, DenseMatrix<Object>, Function1<Object, Object>) - Static method in class org.apache.spark.ml.ann.ApplyInPlace
-
- apply(DenseMatrix<Object>, DenseMatrix<Object>, DenseMatrix<Object>, Function2<Object, Object, Object>) - Static method in class org.apache.spark.ml.ann.ApplyInPlace
-
- apply(String) - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its name.
- apply(int) - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its index.
- apply(T1, T2) - Static method in class org.apache.spark.ml.clustering.ClusterData
-
- apply(T1, T2) - Static method in class org.apache.spark.ml.feature.LabeledPoint
-
- apply(int, int) - Method in class org.apache.spark.ml.linalg.DenseMatrix
-
- apply(int) - Method in class org.apache.spark.ml.linalg.DenseVector
-
- apply(int, int) - Method in interface org.apache.spark.ml.linalg.Matrix
-
Gets the (i, j)-th element.
- apply(int, int) - Method in class org.apache.spark.ml.linalg.SparseMatrix
-
- apply(int) - Method in class org.apache.spark.ml.linalg.SparseVector
-
- apply(int) - Method in interface org.apache.spark.ml.linalg.Vector
-
Gets the value of the ith element.
- apply(Param<T>) - Method in class org.apache.spark.ml.param.ParamMap
-
Gets the value of the input param or its default value if it does not exist.
- apply(GeneralizedLinearRegressionBase) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
-
Constructs the FamilyAndLink object from a parameter map
- apply(T1) - Static method in class org.apache.spark.ml.SaveInstanceEnd
-
- apply(T1) - Static method in class org.apache.spark.ml.SaveInstanceStart
-
- apply() - Static method in class org.apache.spark.ml.TransformEnd
-
- apply() - Static method in class org.apache.spark.ml.TransformStart
-
- apply(Split) - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- apply(Row) - Method in class org.apache.spark.mllib.clustering.KMeansModel.Cluster$
-
- apply(BinaryConfusionMatrix) - Method in interface org.apache.spark.mllib.evaluation.binary.BinaryClassificationMetricComputer
-
- apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.FalsePositiveRate
-
- apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.Precision
-
- apply(BinaryConfusionMatrix) - Static method in class org.apache.spark.mllib.evaluation.binary.Recall
-
- apply(T1) - Static method in class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
-
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.mllib.feature.VocabWord
-
- apply(int, int) - Method in class org.apache.spark.mllib.linalg.DenseMatrix
-
- apply(int) - Method in class org.apache.spark.mllib.linalg.DenseVector
-
- apply(T1, T2) - Static method in class org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- apply(int, int) - Method in interface org.apache.spark.mllib.linalg.Matrix
-
Gets the (i, j)-th element.
- apply(int, int) - Method in class org.apache.spark.mllib.linalg.SparseMatrix
-
- apply(int) - Method in class org.apache.spark.mllib.linalg.SparseVector
-
- apply(int) - Method in interface org.apache.spark.mllib.linalg.Vector
-
Gets the value of the ith element.
- apply(T1, T2, T3) - Static method in class org.apache.spark.mllib.recommendation.Rating
-
- apply(T1, T2) - Static method in class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$.Data
-
- apply(T1, T2) - Static method in class org.apache.spark.mllib.stat.test.BinarySample
-
- apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.Algo
-
- apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
-
- apply(int) - Static method in class org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- apply(int, Node) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData$
-
- apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData$
-
- apply(int, Node) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- apply(Predict) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData$
-
- apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData$
-
- apply(Predict) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- apply(Split) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData$
-
- apply(Row) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData$
-
- apply(Split) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- apply(Row) - Static method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- apply(int, Predict, double, boolean) - Static method in class org.apache.spark.mllib.tree.model.Node
-
Construct a node with nodeIndex, predict, impurity and isLeaf parameters.
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.mllib.tree.model.Split
-
- apply(int) - Static method in class org.apache.spark.rdd.CheckpointState
-
- apply(int) - Static method in class org.apache.spark.rdd.DeterministicLevel
-
- apply(int) - Static method in class org.apache.spark.RequestMethod
-
- apply(T1, T2) - Static method in class org.apache.spark.resource.ResourceInformationJson
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.AccumulableInfo
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- apply(String, long, Enumeration.Value, ByteBuffer, int, Map<String, ResourceInformation>) - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
-
Alternate factory method that takes a ByteBuffer directly for the data field
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.ExcludedExecutor
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.local.KillTask
-
- apply() - Static method in class org.apache.spark.scheduler.local.ReviveOffers
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.local.StatusUpdate
-
- apply() - Static method in class org.apache.spark.scheduler.local.StopExecutor
-
- apply(long, TaskMetrics) - Static method in class org.apache.spark.scheduler.RuntimePercentage
-
- apply(int) - Static method in class org.apache.spark.scheduler.SchedulingMode
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
Deprecated.
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
-
Deprecated.
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorExcluded
-
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
Deprecated.
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerExecutorUnexcluded
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerJobEnd
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.SparkListenerJobStart
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerLogStart
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerMiscellaneousProcessAdded
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
Deprecated.
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
-
Deprecated.
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerNodeExcluded
-
- apply(T1, T2, T3, T4, T5) - Static method in class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
Deprecated.
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerNodeUnexcluded
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerResourceProfileAdded
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerStageCompleted
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.scheduler.SparkListenerTaskEnd
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.scheduler.SparkListenerTaskStart
-
- apply(T1) - Static method in class org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetAdded
-
- apply(T1, T2) - Static method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetRemoved
-
- apply(int) - Static method in class org.apache.spark.scheduler.TaskLocality
-
- apply(Object) - Method in class org.apache.spark.sql.Column
-
Extracts a value or values from a complex type.
- apply(String, Expression...) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
-
Create a logical transform for applying a named transform.
- apply(String, Seq<Expression>) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
-
- apply(String) - Method in class org.apache.spark.sql.Dataset
-
Selects column based on the column name and returns it as a
Column
.
- apply(Column...) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Deprecated.
Creates a Column
for this UDAF using given Column
s as input arguments.
- apply(Seq<Column>) - Method in class org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Deprecated.
Creates a Column
for this UDAF using given Column
s as input arguments.
- apply(Column...) - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
-
Returns an expression that invokes the UDF, using the given arguments.
- apply(Seq<Column>) - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
-
Returns an expression that invokes the UDF, using the given arguments.
- apply(T1, T2) - Static method in class org.apache.spark.sql.jdbc.JdbcType
-
- apply() - Static method in class org.apache.spark.sql.Observation
-
Observation constructor for creating an anonymous observation.
- apply(String) - Static method in class org.apache.spark.sql.Observation
-
Observation constructor for creating a named observation.
- apply(Dataset<Row>, Seq<Expression>, RelationalGroupedDataset.GroupType) - Static method in class org.apache.spark.sql.RelationalGroupedDataset
-
- apply(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i.
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.And
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.EqualNullSafe
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.EqualTo
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.GreaterThan
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.GreaterThanOrEqual
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.In
-
- apply(T1) - Static method in class org.apache.spark.sql.sources.IsNotNull
-
- apply(T1) - Static method in class org.apache.spark.sql.sources.IsNull
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.LessThan
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.LessThanOrEqual
-
- apply(T1) - Static method in class org.apache.spark.sql.sources.Not
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.Or
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringContains
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringEndsWith
-
- apply(T1, T2) - Static method in class org.apache.spark.sql.sources.StringStartsWith
-
- apply(String, Option<Object>, Map<String, String>) - Static method in class org.apache.spark.sql.streaming.SinkProgress
-
- apply(DataType) - Static method in class org.apache.spark.sql.types.ArrayType
-
Construct a
ArrayType
object with the given element type.
- apply(T1) - Static method in class org.apache.spark.sql.types.CharType
-
- apply() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- apply(byte) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- apply(double) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(long) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(int) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigInteger) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigInt) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal, int, int) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal, int, int) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(long, int, int) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(String) - Static method in class org.apache.spark.sql.types.Decimal
-
- apply(DataType, DataType) - Static method in class org.apache.spark.sql.types.MapType
-
Construct a
MapType
object with the given key type and value type.
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.sql.types.StructField
-
- apply(String) - Method in class org.apache.spark.sql.types.StructType
-
- apply(Set<String>) - Method in class org.apache.spark.sql.types.StructType
-
Returns a
StructType
containing
StructField
s of the given names, preserving the
original order of fields.
- apply(int) - Method in class org.apache.spark.sql.types.StructType
-
- apply(T1) - Static method in class org.apache.spark.sql.types.VarcharType
-
- apply() - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
-
- apply(byte) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
-
- apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.status.api.v1.ApplicationInfo
-
- apply(T1, T2) - Static method in class org.apache.spark.status.api.v1.sql.Metric
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.status.api.v1.sql.Node
-
- apply(T1) - Static method in class org.apache.spark.status.api.v1.StackTrace
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.status.api.v1.ThreadStackTrace
-
- apply(int) - Method in class org.apache.spark.status.RDDPartitionSeq
-
- apply(String) - Static method in class org.apache.spark.storage.BlockId
-
- apply(String, String, int, Option<String>) - Static method in class org.apache.spark.storage.BlockManagerId
-
- apply(ObjectInput) - Static method in class org.apache.spark.storage.BlockManagerId
-
- apply(T1, T2) - Static method in class org.apache.spark.storage.BroadcastBlockId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.CacheId
-
- apply(T1, T2) - Static method in class org.apache.spark.storage.RDDBlockId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleBlockBatchId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleBlockChunkId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleBlockId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleChecksumBlockId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleDataBlockId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleIndexBlockId
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.storage.ShuffleMergedBlockId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedDataBlockId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.storage.ShufflePushBlockId
-
- apply(boolean, boolean, boolean, boolean, int) - Static method in class org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object.
- apply(boolean, boolean, boolean, int) - Static method in class org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object without setting useOffHeap.
- apply(int, int) - Static method in class org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object from its integer representation.
- apply(ObjectInput) - Static method in class org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Read StorageLevel object from ObjectInput stream.
- apply(T1, T2) - Static method in class org.apache.spark.storage.StreamBlockId
-
- apply(T1) - Static method in class org.apache.spark.storage.TaskResultBlockId
-
- apply(T1) - Static method in class org.apache.spark.streaming.Duration
-
- apply(long) - Static method in class org.apache.spark.streaming.Milliseconds
-
- apply(long) - Static method in class org.apache.spark.streaming.Minutes
-
- apply(T1, T2, T3, T4, T5, T6) - Static method in class org.apache.spark.streaming.scheduler.BatchInfo
-
- apply(T1, T2, T3, T4, T5, T6, T7) - Static method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- apply(T1, T2, T3, T4, T5, T6, T7, T8) - Static method in class org.apache.spark.streaming.scheduler.ReceiverInfo
-
- apply(int) - Static method in class org.apache.spark.streaming.scheduler.ReceiverState
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- apply(T1) - Static method in class org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- apply(long) - Static method in class org.apache.spark.streaming.Seconds
-
- apply(T1, T2, T3) - Static method in class org.apache.spark.TaskCommitDenied
-
- apply(T1, T2, T3, T4) - Static method in class org.apache.spark.TaskKilled
-
- apply(int) - Static method in class org.apache.spark.TaskState
-
- apply(TraversableOnce<Object>) - Static method in class org.apache.spark.util.StatCounter
-
Build a StatCounter from a list of values.
- apply(Seq<Object>) - Static method in class org.apache.spark.util.StatCounter
-
Build a StatCounter from a list of values passed as variable-length arguments.
- ApplyInPlace - Class in org.apache.spark.ml.ann
-
Implements in-place application of functions in the arrays
- ApplyInPlace() - Constructor for class org.apache.spark.ml.ann.ApplyInPlace
-
- applyNamespaceChanges(Map<String, String>, Seq<NamespaceChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Apply properties changes to a map and return the result.
- applyNamespaceChanges(Map<String, String>, Seq<NamespaceChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Apply properties changes to a Java map and return the result.
- applyPropertiesChanges(Map<String, String>, Seq<TableChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Apply properties changes to a map and return the result.
- applyPropertiesChanges(Map<String, String>, Seq<TableChange>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Apply properties changes to a Java map and return the result.
- applySchema(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
-
- applySchema(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
-
- applySchema(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
-
- applySchema(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
-
- applySchemaChanges(StructType, Seq<TableChange>, Option<String>, String) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Apply schema changes to a schema and return the result.
- appName() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- appName() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- appName() - Method in class org.apache.spark.SparkContext
-
- appName(String) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a name for the application, which will be shown in the Spark web UI.
- approx_count_distinct(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(Column, double) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(String, double) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_percentile(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the approximate percentile
of the numeric column col
which
is the smallest value in the ordered col
values (sorted from least to greatest) such that
no more than percentage
of col
values is less than the value or equal to that value.
- approxCountDistinct(Column) - Static method in class org.apache.spark.sql.functions
-
- approxCountDistinct(String) - Static method in class org.apache.spark.sql.functions
-
- approxCountDistinct(Column, double) - Static method in class org.apache.spark.sql.functions
-
- approxCountDistinct(String, double) - Static method in class org.apache.spark.sql.functions
-
- ApproxHist() - Static method in class org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- ApproximateEvaluator<U,R> - Interface in org.apache.spark.partial
-
An object that computes a function incrementally by merging in results of type U from multiple
tasks.
- approxQuantile(String, double[], double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the approximate quantiles of a numerical column of a DataFrame.
- approxQuantile(String[], double[], double) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the approximate quantiles of numerical columns of a DataFrame.
- appSparkVersion() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- AppStatusUtils - Class in org.apache.spark.status
-
- AppStatusUtils() - Constructor for class org.apache.spark.status.AppStatusUtils
-
- archives() - Method in class org.apache.spark.SparkContext
-
- AreaUnderCurve - Class in org.apache.spark.mllib.evaluation
-
Computes the area under the curve (AUC) using the trapezoidal rule.
- AreaUnderCurve() - Constructor for class org.apache.spark.mllib.evaluation.AreaUnderCurve
-
- areaUnderPR() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Computes the area under the precision-recall curve.
- areaUnderROC() - Method in interface org.apache.spark.ml.classification.BinaryClassificationSummary
-
Computes the area under the receiver operating characteristic (ROC) curve.
- areaUnderROC() - Method in class org.apache.spark.ml.classification.BinaryLogisticRegressionSummaryImpl
-
- areaUnderROC() - Method in class org.apache.spark.ml.classification.BinaryRandomForestClassificationSummaryImpl
-
- areaUnderROC() - Method in class org.apache.spark.ml.classification.FMClassificationSummaryImpl
-
- areaUnderROC() - Method in class org.apache.spark.ml.classification.LinearSVCSummaryImpl
-
- areaUnderROC() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Computes the area under the receiver operating characteristic (ROC) curve.
- argmax() - Method in class org.apache.spark.ml.linalg.DenseVector
-
- argmax() - Method in class org.apache.spark.ml.linalg.SparseVector
-
- argmax() - Method in interface org.apache.spark.ml.linalg.Vector
-
Find the index of a maximal element.
- argmax() - Method in class org.apache.spark.mllib.linalg.DenseVector
-
- argmax() - Method in class org.apache.spark.mllib.linalg.SparseVector
-
- argmax() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Find the index of a maximal element.
- arguments() - Method in interface org.apache.spark.sql.connector.expressions.Transform
-
Returns the arguments passed to the transform function.
- arithmeticOverflowError(String, String, SQLQueryContext) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- arithmeticOverflowError(ArithmeticException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- arithmeticOverflowError$default$2() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- arithmeticOverflowError$default$3() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- ARPACK - Class in org.apache.spark.mllib.linalg
-
ARPACK routines for MLlib's vectors and matrices.
- ARPACK() - Constructor for class org.apache.spark.mllib.linalg.ARPACK
-
- array(DataType) - Method in class org.apache.spark.sql.ColumnName
-
Creates a new StructField
of type array.
- array(Column...) - Static method in class org.apache.spark.sql.functions
-
Creates a new array column.
- array(String, String...) - Static method in class org.apache.spark.sql.functions
-
Creates a new array column.
- array(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Creates a new array column.
- array(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
-
Creates a new array column.
- array() - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- array_agg(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns a list of objects with duplicates.
- array_append(Column, Object) - Static method in class org.apache.spark.sql.functions
-
Returns an ARRAY containing all elements from the source ARRAY as well as the new element.
- array_compact(Column) - Static method in class org.apache.spark.sql.functions
-
Remove all null elements from the given array.
- array_contains(Column, Object) - Static method in class org.apache.spark.sql.functions
-
Returns null if the array is null, true if the array contains value
, and false otherwise.
- array_distinct(Column) - Static method in class org.apache.spark.sql.functions
-
Removes duplicate values from the array.
- array_except(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an array of the elements in the first array but not in the second array,
without duplicates.
- array_insert(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Adds an item into a given array at a specified position
- array_intersect(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an array of the elements in the intersection of the given two arrays,
without duplicates.
- array_join(Column, String, String) - Static method in class org.apache.spark.sql.functions
-
Concatenates the elements of column
using the delimiter
.
- array_join(Column, String) - Static method in class org.apache.spark.sql.functions
-
Concatenates the elements of column
using the delimiter
.
- array_max(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the maximum value in the array.
- array_min(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the minimum value in the array.
- array_position(Column, Object) - Static method in class org.apache.spark.sql.functions
-
Locates the position of the first occurrence of the value in the given array as long.
- array_prepend(Column, Object) - Static method in class org.apache.spark.sql.functions
-
Returns an array containing value as well as all elements from array.
- array_remove(Column, Object) - Static method in class org.apache.spark.sql.functions
-
Remove all elements that equal to element from the given array.
- array_repeat(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Creates an array containing the left argument repeated the number of times given by the
right argument.
- array_repeat(Column, int) - Static method in class org.apache.spark.sql.functions
-
Creates an array containing the left argument repeated the number of times given by the
right argument.
- array_size(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the total number of elements in the array.
- array_sort(Column) - Static method in class org.apache.spark.sql.functions
-
Sorts the input array in ascending order.
- array_sort(Column, Function2<Column, Column, Column>) - Static method in class org.apache.spark.sql.functions
-
Sorts the input array based on the given comparator function.
- array_to_vector(Column) - Static method in class org.apache.spark.ml.functions
-
Converts a column of array of numeric type into a column of dense vectors in MLlib.
- array_union(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an array of the elements in the union of the given two arrays, without duplicates.
- arrayComponentTypeUnsupportedError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- arrayLengthGt(double) - Static method in class org.apache.spark.ml.param.ParamValidators
-
Check that the array length is greater than lowerBound.
- arrays_overlap(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns true
if a1
and a2
have at least one non-null element in common.
- arrays_zip(Column...) - Static method in class org.apache.spark.sql.functions
-
Returns a merged array of structs in which the N-th struct contains all N-th values of input
arrays.
- arrays_zip(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Returns a merged array of structs in which the N-th struct contains all N-th values of input
arrays.
- ArrayType - Class in org.apache.spark.sql.types
-
- ArrayType(DataType, boolean) - Constructor for class org.apache.spark.sql.types.ArrayType
-
- arrayValues() - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
-
- ArrowColumnVector - Class in org.apache.spark.sql.vectorized
-
A column vector backed by Apache Arrow.
- ArrowColumnVector(ValueVector) - Constructor for class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- ArrowUtils - Class in org.apache.spark.sql.util
-
- ArrowUtils() - Constructor for class org.apache.spark.sql.util.ArrowUtils
-
- as(Encoder<U>) - Method in class org.apache.spark.sql.Column
-
Provides a type hint about the expected return value of this column.
- as(String) - Method in class org.apache.spark.sql.Column
-
Gives the column an alias.
- as(Seq<String>) - Method in class org.apache.spark.sql.Column
-
(Scala-specific) Assigns the given aliases to the results of a table generating function.
- as(String[]) - Method in class org.apache.spark.sql.Column
-
Assigns the given aliases to the results of a table generating function.
- as(Symbol) - Method in class org.apache.spark.sql.Column
-
Gives the column an alias.
- as(String, Metadata) - Method in class org.apache.spark.sql.Column
-
Gives the column an alias with metadata.
- as(Encoder<U>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset where each record has been mapped on to the specified type.
- as(String) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with an alias set.
- as(Symbol) - Method in class org.apache.spark.sql.Dataset
-
(Scala-specific) Returns a new Dataset with an alias set.
- as(Encoder<K>, Encoder<T>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Returns a KeyValueGroupedDataset
where the data is grouped by the grouping expressions
of current RelationalGroupedDataset
.
- asBinary() - Method in interface org.apache.spark.ml.classification.LogisticRegressionSummary
-
Convenient method for casting to binary logistic regression summary.
- asBinary() - Method in interface org.apache.spark.ml.classification.RandomForestClassificationSummary
-
Convenient method for casting to BinaryRandomForestClassificationSummary.
- asBreeze() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Converts to a breeze matrix.
- asBreeze() - Method in interface org.apache.spark.ml.linalg.Vector
-
Converts the instance to a breeze vector.
- asBreeze() - Method in interface org.apache.spark.mllib.linalg.Matrix
-
Converts to a breeze matrix.
- asBreeze() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Converts the instance to a breeze vector.
- asc() - Method in class org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column.
- asc(String) - Static method in class org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column.
- asc_nulls_first() - Method in class org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column,
and null values return before non-null values.
- asc_nulls_first(String) - Static method in class org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column,
and null values return before non-null values.
- asc_nulls_last() - Method in class org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column,
and null values appear after non-null values.
- asc_nulls_last(String) - Static method in class org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column,
and null values appear after non-null values.
- asCaseSensitiveMap() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
Returns the original case-sensitive map.
- ascii(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the numeric value of the first character of the string column, and returns the
result as an int column.
- asFunctionCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
-
- asFunctionIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
-
- asFunctionIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
-
- asIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
-
- asin(Column) - Static method in class org.apache.spark.sql.functions
-
- asin(String) - Static method in class org.apache.spark.sql.functions
-
- asinh(Column) - Static method in class org.apache.spark.sql.functions
-
- asinh(String) - Static method in class org.apache.spark.sql.functions
-
- asInteraction() - Static method in class org.apache.spark.ml.feature.Dot
-
- asInteraction() - Method in interface org.apache.spark.ml.feature.InteractableTerm
-
Convert to ColumnInteraction to wrap all interactions.
- asIterator() - Method in class org.apache.spark.serializer.DeserializationStream
-
Read the elements of this stream through an iterator.
- asJavaPairRDD() - Method in class org.apache.spark.api.r.PairwiseRRDD
-
- asJavaRDD() - Method in class org.apache.spark.api.r.RRDD
-
- asJavaRDD() - Method in class org.apache.spark.api.r.StringRRDD
-
- ask(Object) - Method in interface org.apache.spark.api.plugin.PluginContext
-
Send an RPC to the plugin's driver-side component.
- asKeyValueIterator() - Method in class org.apache.spark.serializer.DeserializationStream
-
Read the elements of this stream through an iterator over key-value pairs.
- AskPermissionToCommitOutput - Class in org.apache.spark.scheduler
-
- AskPermissionToCommitOutput(int, int, int, int) - Constructor for class org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- askRpcTimeout(SparkConf) - Static method in class org.apache.spark.util.RpcUtils
-
Returns the default Spark timeout to use for RPC ask operations.
- askStandaloneSchedulerToShutDownExecutorsError(Exception) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- askStorageEndpoints() - Method in class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
-
- askStorageEndpoints() - Method in class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
-
- asML() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
-
- asML() - Method in class org.apache.spark.mllib.linalg.DenseVector
-
- asML() - Method in interface org.apache.spark.mllib.linalg.Matrix
-
Convert this matrix to the new mllib-local representation.
- asML() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
-
- asML() - Method in class org.apache.spark.mllib.linalg.SparseVector
-
- asML() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Convert this vector to the new mllib-local representation.
- asMultipart() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.FunctionIdentifierHelper
-
- asMultipartIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
-
- asNamespaceCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
-
- asNondeterministic() - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
-
Updates UserDefinedFunction to nondeterministic.
- asNonNullable() - Method in class org.apache.spark.sql.expressions.UserDefinedFunction
-
Updates UserDefinedFunction to non-nullable.
- asNullable() - Method in class org.apache.spark.sql.types.ObjectType
-
- asRDDId() - Method in class org.apache.spark.storage.BlockId
-
- asSchema() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.ColumnsHelper
-
- assert_true(Column) - Static method in class org.apache.spark.sql.functions
-
Returns null if the condition is true, and throws an exception otherwise.
- assert_true(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns null if the condition is true; throws an exception with the error message otherwise.
- assertExceptionMsg(Throwable, String, boolean, ClassTag<E>) - Static method in class org.apache.spark.TestUtils
-
Asserts that exception message contains the message.
- assertNotSpilled(SparkContext, String, Function0<BoxedUnit>) - Static method in class org.apache.spark.TestUtils
-
Run some code involving jobs submitted to the given context and assert that the jobs
did not spill.
- assertSpilled(SparkContext, String, Function0<BoxedUnit>) - Static method in class org.apache.spark.TestUtils
-
Run some code involving jobs submitted to the given context and assert that the jobs spilled.
- assignClusters(Dataset<?>) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
Run the PIC algorithm and returns a cluster assignment for each input vertex.
- assignedAddrs() - Method in interface org.apache.spark.resource.ResourceAllocator
-
Sequence of currently assigned resource addresses.
- Assignment(long, int) - Constructor for class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
-
- Assignment$() - Constructor for class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment$
-
- assignments() - Method in class org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- AssociationRules - Class in org.apache.spark.ml.fpm
-
- AssociationRules() - Constructor for class org.apache.spark.ml.fpm.AssociationRules
-
- associationRules() - Method in class org.apache.spark.ml.fpm.FPGrowthModel
-
Get association rules fitted using the minConfidence.
- AssociationRules - Class in org.apache.spark.mllib.fpm
-
Generates association rules from a RDD[FreqItemset[Item}
.
- AssociationRules() - Constructor for class org.apache.spark.mllib.fpm.AssociationRules
-
Constructs a default instance with default parameters {minConfidence = 0.8}.
- AssociationRules.Rule<Item> - Class in org.apache.spark.mllib.fpm
-
An association rule between sets of items.
- asTableCatalog() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
-
- asTableIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.IdentifierHelper
-
- asTableIdentifier() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.MultipartIdentifierHelper
-
- AsTableIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- AsTableIdentifier() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier
-
- AsTableIdentifier$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.AsTableIdentifier$
-
- asTerms() - Static method in class org.apache.spark.ml.feature.Dot
-
- asTerms() - Static method in class org.apache.spark.ml.feature.EmptyTerm
-
- asTerms() - Method in interface org.apache.spark.ml.feature.Term
-
Default representation of a single Term as a part of summed terms.
- asTransform() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.BucketSpecHelper
-
- asTransforms() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.PartitionTypeHelper
-
- AsyncEventQueue - Class in org.apache.spark.scheduler
-
An asynchronous queue for events.
- AsyncEventQueue(String, SparkConf, LiveListenerBusMetrics, LiveListenerBus) - Constructor for class org.apache.spark.scheduler.AsyncEventQueue
-
- AsyncRDDActions<T> - Class in org.apache.spark.rdd
-
A set of asynchronous RDD actions available through an implicit conversion.
- AsyncRDDActions(RDD<T>, ClassTag<T>) - Constructor for class org.apache.spark.rdd.AsyncRDDActions
-
- atan(Column) - Static method in class org.apache.spark.sql.functions
-
- atan(String) - Static method in class org.apache.spark.sql.functions
-
- atan2(Column, Column) - Static method in class org.apache.spark.sql.functions
-
- atan2(Column, String) - Static method in class org.apache.spark.sql.functions
-
- atan2(String, Column) - Static method in class org.apache.spark.sql.functions
-
- atan2(String, String) - Static method in class org.apache.spark.sql.functions
-
- atan2(Column, double) - Static method in class org.apache.spark.sql.functions
-
- atan2(String, double) - Static method in class org.apache.spark.sql.functions
-
- atan2(double, Column) - Static method in class org.apache.spark.sql.functions
-
- atan2(double, String) - Static method in class org.apache.spark.sql.functions
-
- atanh(Column) - Static method in class org.apache.spark.sql.functions
-
- atanh(String) - Static method in class org.apache.spark.sql.functions
-
- attempt() - Method in class org.apache.spark.status.api.v1.TaskData
-
- ATTEMPT() - Static method in class org.apache.spark.status.TaskIndexNames
-
- ATTEMPT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- ATTEMPT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- attemptId() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- attemptId() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
-
- attemptId() - Method in class org.apache.spark.status.api.v1.StageData
-
- attemptNumber() - Method in class org.apache.spark.BarrierTaskContext
-
- attemptNumber() - Method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- attemptNumber() - Method in class org.apache.spark.scheduler.StageInfo
-
- attemptNumber() - Method in class org.apache.spark.scheduler.TaskInfo
-
- attemptNumber() - Method in class org.apache.spark.TaskCommitDenied
-
- attemptNumber() - Method in class org.apache.spark.TaskContext
-
How many times this task has been attempted.
- attempts() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
-
- ATTEMPTS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- AtTimestamp(Date) - Constructor for class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- attr() - Method in class org.apache.spark.graphx.Edge
-
- attr() - Method in class org.apache.spark.graphx.EdgeContext
-
The attribute associated with the edge.
- attr() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- Attribute - Class in org.apache.spark.ml.attribute
-
Abstract class for ML attributes.
- Attribute() - Constructor for class org.apache.spark.ml.attribute.Attribute
-
- attribute() - Method in class org.apache.spark.sql.sources.EqualNullSafe
-
- attribute() - Method in class org.apache.spark.sql.sources.EqualTo
-
- attribute() - Method in class org.apache.spark.sql.sources.GreaterThan
-
- attribute() - Method in class org.apache.spark.sql.sources.GreaterThanOrEqual
-
- attribute() - Method in class org.apache.spark.sql.sources.In
-
- attribute() - Method in class org.apache.spark.sql.sources.IsNotNull
-
- attribute() - Method in class org.apache.spark.sql.sources.IsNull
-
- attribute() - Method in class org.apache.spark.sql.sources.LessThan
-
- attribute() - Method in class org.apache.spark.sql.sources.LessThanOrEqual
-
- attribute() - Method in class org.apache.spark.sql.sources.StringContains
-
- attribute() - Method in class org.apache.spark.sql.sources.StringEndsWith
-
- attribute() - Method in class org.apache.spark.sql.sources.StringStartsWith
-
- AttributeFactory - Interface in org.apache.spark.ml.attribute
-
Trait for ML attribute factories.
- AttributeGroup - Class in org.apache.spark.ml.attribute
-
Attributes that describe a vector ML column.
- AttributeGroup(String) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group without attribute info.
- AttributeGroup(String, int) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group knowing only the number of attributes.
- AttributeGroup(String, Attribute[]) - Constructor for class org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group with attributes.
- AttributeKeys - Class in org.apache.spark.ml.attribute
-
Keys used to store attributes.
- AttributeKeys() - Constructor for class org.apache.spark.ml.attribute.AttributeKeys
-
- attributeNameSyntaxError(String) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- attributeNameSyntaxError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- attributes() - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Optional array of attributes.
- ATTRIBUTES() - Static method in class org.apache.spark.ml.attribute.AttributeKeys
-
- attributes() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
-
- attributes() - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
-
- attributes() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
-
- ATTRIBUTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- attributesForTypeUnsupportedError(ScalaReflection.Schema) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- AttributeType - Class in org.apache.spark.ml.attribute
-
An enum-like type for attribute types: AttributeType$.Numeric
, AttributeType$.Nominal
,
and AttributeType$.Binary
.
- AttributeType(String) - Constructor for class org.apache.spark.ml.attribute.AttributeType
-
- attrType() - Method in class org.apache.spark.ml.attribute.Attribute
-
Attribute type.
- attrType() - Method in class org.apache.spark.ml.attribute.BinaryAttribute
-
- attrType() - Method in class org.apache.spark.ml.attribute.NominalAttribute
-
- attrType() - Method in class org.apache.spark.ml.attribute.NumericAttribute
-
- attrType() - Static method in class org.apache.spark.ml.attribute.UnresolvedAttribute
-
- available() - Method in class org.apache.spark.io.NioBufferedFileInputStream
-
- available() - Method in class org.apache.spark.io.ReadAheadInputStream
-
- available() - Method in class org.apache.spark.storage.BufferReleasingInputStream
-
- availableAddrs() - Method in interface org.apache.spark.resource.ResourceAllocator
-
Sequence of currently available resource addresses.
- AvailableNow() - Static method in class org.apache.spark.sql.streaming.Trigger
-
A trigger that processes all available data at the start of the query in one or multiple
batches, then terminates the query.
- Average() - Static method in class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- Avg - Class in org.apache.spark.sql.connector.expressions.aggregate
-
An aggregate function that returns the mean of all the values in a group.
- Avg(Expression, boolean) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Avg
-
- avg(MapFunction<T, Double>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
-
Deprecated.
Average aggregate function.
- avg(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
-
Deprecated.
Average aggregate function.
- avg(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- avg(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- avg(String...) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute the mean value for each numeric columns for each group.
- avg(Seq<String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute the mean value for each numeric columns for each group.
- avg() - Method in class org.apache.spark.util.DoubleAccumulator
-
Returns the average of elements added to the accumulator.
- avg() - Method in class org.apache.spark.util.LongAccumulator
-
Returns the average of elements added to the accumulator.
- avgEventRate() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- avgInputRate() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgLen() - Method in interface org.apache.spark.sql.connector.read.colstats.ColumnStatistics
-
- avgMetrics() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
-
- avgProcessingTime() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgSchedulingDelay() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgTotalDelay() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avroIncompatibleReadError(String, String, String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- AvroMatchedField$() - Constructor for class org.apache.spark.sql.avro.AvroUtils.AvroMatchedField$
-
- AvroSchemaHelper(Schema, StructType, Seq<String>, Seq<String>, boolean) - Constructor for class org.apache.spark.sql.avro.AvroUtils.AvroSchemaHelper
-
- AvroUtils - Class in org.apache.spark.sql.avro
-
- AvroUtils() - Constructor for class org.apache.spark.sql.avro.AvroUtils
-
- AvroUtils.AvroMatchedField$ - Class in org.apache.spark.sql.avro
-
- AvroUtils.AvroSchemaHelper - Class in org.apache.spark.sql.avro
-
Helper class to perform field lookup/matching on Avro schemas.
- AvroUtils.RowReader - Interface in org.apache.spark.sql.avro
-
- awaitAnyTermination() - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
Wait until any of the queries on the associated SQLContext has terminated since the
creation of the context, or since resetTerminated()
was called.
- awaitAnyTermination(long) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
Wait until any of the queries on the associated SQLContext has terminated since the
creation of the context, or since resetTerminated()
was called.
- awaitReady(Awaitable<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
-
Preferred alternative to Await.ready()
.
- awaitResult(Awaitable<T>, Duration) - Static method in class org.apache.spark.util.SparkThreadUtils
-
Preferred alternative to Await.result()
.
- awaitResult(Awaitable<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
-
Preferred alternative to Await.result()
.
- awaitResult(Future<T>, Duration) - Static method in class org.apache.spark.util.ThreadUtils
-
- awaitTermination() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
-
Waits for the termination of this
query, either by query.stop()
or by an exception.
- awaitTermination(long) - Method in interface org.apache.spark.sql.streaming.StreamingQuery
-
Waits for the termination of this
query, either by query.stop()
or by an exception.
- awaitTermination() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Wait for the execution to stop.
- awaitTermination() - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Wait for the execution to stop.
- awaitTerminationOrTimeout(long) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Wait for the execution to stop.
- awaitTerminationOrTimeout(long) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Wait for the execution to stop.
- axpy(double, Vector, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
-
y += a * x
- axpy(double, Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
y += a * x
- cache() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY
).
- cache() - Method in class org.apache.spark.api.java.JavaPairRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY
).
- cache() - Method in class org.apache.spark.api.java.JavaRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY
).
- cache() - Method in class org.apache.spark.graphx.Graph
-
Caches the vertices and edges associated with this graph at the previously-specified target
storage levels, which default to MEMORY_ONLY
.
- cache() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
Persists the edge partitions using targetStorageLevel
, which defaults to MEMORY_ONLY.
- cache() - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- cache() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
Persists the vertex partitions at targetStorageLevel
, which defaults to MEMORY_ONLY.
- cache() - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Caches the underlying RDD.
- cache() - Method in class org.apache.spark.rdd.RDD
-
Persist this RDD with the default storage level (MEMORY_ONLY
).
- cache() - Method in class org.apache.spark.sql.Dataset
-
Persist this Dataset with the default storage level (MEMORY_AND_DISK
).
- cache() - Method in class org.apache.spark.streaming.api.java.JavaDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- cache() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- cache() - Method in class org.apache.spark.streaming.dstream.DStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- CACHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- CACHED_PARTITIONS() - Static method in class org.apache.spark.ui.storage.ToolTips
-
- CachedBatch - Interface in org.apache.spark.sql.columnar
-
Basic interface that all cached batches of data must support.
- CachedBatchSerializer - Interface in org.apache.spark.sql.columnar
-
Provides APIs that handle transformations of SQL data associated with the cache/persist APIs.
- CacheId - Class in org.apache.spark.storage
-
- CacheId(String, String, String) - Constructor for class org.apache.spark.storage.CacheId
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
-
- cacheNodeIds() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- cacheNodeIds() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
If false, the algorithm will pass trees to executors to match instances with nodes.
- cacheSize() - Method in interface org.apache.spark.SparkExecutorInfo
-
- cacheSize() - Method in class org.apache.spark.SparkExecutorInfoImpl
-
- cacheTable(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Caches the specified table in-memory.
- cacheTable(String, StorageLevel) - Method in class org.apache.spark.sql.catalog.Catalog
-
Caches the specified table with the given storage level.
- cacheTable(String) - Method in class org.apache.spark.sql.SQLContext
-
Caches the specified table in-memory.
- calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Entropy
-
information calculation for multiclass classification
- calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Entropy
-
variance calculation
- calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Gini
-
information calculation for multiclass classification
- calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Gini
-
variance calculation
- calculate(double[], double) - Method in interface org.apache.spark.mllib.tree.impurity.Impurity
-
information calculation for multiclass classification
- calculate(double, double, double) - Method in interface org.apache.spark.mllib.tree.impurity.Impurity
-
information calculation for regression
- calculate(double[], double) - Static method in class org.apache.spark.mllib.tree.impurity.Variance
-
information calculation for multiclass classification
- calculate(double, double, double) - Static method in class org.apache.spark.mllib.tree.impurity.Variance
-
variance calculation
- calculateAmountAndPartsForFraction(double) - Static method in class org.apache.spark.resource.ResourceUtils
-
- calculateNumberOfPartitions(long, int, int) - Method in class org.apache.spark.ml.feature.Word2VecModel.Word2VecModelWriter$
-
Calculate the number of partitions to use in saving the model.
- CalendarInterval - Class in org.apache.spark.unsafe.types
-
The class representing calendar intervals.
- CalendarInterval(int, int, long) - Constructor for class org.apache.spark.unsafe.types.CalendarInterval
-
- CalendarIntervalType - Class in org.apache.spark.sql.types
-
The data type representing calendar intervals.
- CalendarIntervalType() - Constructor for class org.apache.spark.sql.types.CalendarIntervalType
-
- CalendarIntervalType - Static variable in class org.apache.spark.sql.types.DataTypes
-
Gets the CalendarIntervalType object.
- call(K, Iterator<V1>, Iterator<V2>) - Method in interface org.apache.spark.api.java.function.CoGroupFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.DoubleFlatMapFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.DoubleFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.FilterFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.FlatMapFunction
-
- call(T1, T2) - Method in interface org.apache.spark.api.java.function.FlatMapFunction2
-
- call(K, Iterator<V>) - Method in interface org.apache.spark.api.java.function.FlatMapGroupsFunction
-
- call(K, Iterator<V>, GroupState<S>) - Method in interface org.apache.spark.api.java.function.FlatMapGroupsWithStateFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.ForeachFunction
-
- call(Iterator<T>) - Method in interface org.apache.spark.api.java.function.ForeachPartitionFunction
-
- call(T1) - Method in interface org.apache.spark.api.java.function.Function
-
- call() - Method in interface org.apache.spark.api.java.function.Function0
-
- call(T1, T2) - Method in interface org.apache.spark.api.java.function.Function2
-
- call(T1, T2, T3) - Method in interface org.apache.spark.api.java.function.Function3
-
- call(T1, T2, T3, T4) - Method in interface org.apache.spark.api.java.function.Function4
-
- call(T) - Method in interface org.apache.spark.api.java.function.MapFunction
-
- call(K, Iterator<V>) - Method in interface org.apache.spark.api.java.function.MapGroupsFunction
-
- call(K, Iterator<V>, GroupState<S>) - Method in interface org.apache.spark.api.java.function.MapGroupsWithStateFunction
-
- call(Iterator<T>) - Method in interface org.apache.spark.api.java.function.MapPartitionsFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.PairFlatMapFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.PairFunction
-
- call(T, T) - Method in interface org.apache.spark.api.java.function.ReduceFunction
-
- call(T) - Method in interface org.apache.spark.api.java.function.VoidFunction
-
- call(T1, T2) - Method in interface org.apache.spark.api.java.function.VoidFunction2
-
- call() - Method in interface org.apache.spark.sql.api.java.UDF0
-
- call(T1) - Method in interface org.apache.spark.sql.api.java.UDF1
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10) - Method in interface org.apache.spark.sql.api.java.UDF10
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11) - Method in interface org.apache.spark.sql.api.java.UDF11
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12) - Method in interface org.apache.spark.sql.api.java.UDF12
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13) - Method in interface org.apache.spark.sql.api.java.UDF13
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14) - Method in interface org.apache.spark.sql.api.java.UDF14
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15) - Method in interface org.apache.spark.sql.api.java.UDF15
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16) - Method in interface org.apache.spark.sql.api.java.UDF16
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17) - Method in interface org.apache.spark.sql.api.java.UDF17
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18) - Method in interface org.apache.spark.sql.api.java.UDF18
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19) - Method in interface org.apache.spark.sql.api.java.UDF19
-
- call(T1, T2) - Method in interface org.apache.spark.sql.api.java.UDF2
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20) - Method in interface org.apache.spark.sql.api.java.UDF20
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21) - Method in interface org.apache.spark.sql.api.java.UDF21
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22) - Method in interface org.apache.spark.sql.api.java.UDF22
-
- call(T1, T2, T3) - Method in interface org.apache.spark.sql.api.java.UDF3
-
- call(T1, T2, T3, T4) - Method in interface org.apache.spark.sql.api.java.UDF4
-
- call(T1, T2, T3, T4, T5) - Method in interface org.apache.spark.sql.api.java.UDF5
-
- call(T1, T2, T3, T4, T5, T6) - Method in interface org.apache.spark.sql.api.java.UDF6
-
- call(T1, T2, T3, T4, T5, T6, T7) - Method in interface org.apache.spark.sql.api.java.UDF7
-
- call(T1, T2, T3, T4, T5, T6, T7, T8) - Method in interface org.apache.spark.sql.api.java.UDF8
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9) - Method in interface org.apache.spark.sql.api.java.UDF9
-
- call_function(String, Column...) - Static method in class org.apache.spark.sql.functions
-
Call a SQL function.
- call_function(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Call a SQL function.
- call_udf(String, Column...) - Static method in class org.apache.spark.sql.functions
-
Call an user-defined function.
- call_udf(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Call an user-defined function.
- callSite() - Method in class org.apache.spark.storage.RDDInfo
-
- CALLSITE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- callUDF(String, Column...) - Static method in class org.apache.spark.sql.functions
-
Call an user-defined function.
- callUDF(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
- cancel() - Method in class org.apache.spark.ComplexFutureAction
-
- cancel() - Method in interface org.apache.spark.FutureAction
-
Cancels the execution of this action.
- cancel() - Method in class org.apache.spark.SimpleFutureAction
-
- cancelAllJobs() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Cancel all jobs that have been scheduled or are running.
- cancelAllJobs() - Method in class org.apache.spark.SparkContext
-
Cancel all jobs that have been scheduled or are running.
- cancelJob(int, String) - Method in class org.apache.spark.SparkContext
-
Cancel a given job if it's scheduled or running.
- cancelJob(int) - Method in class org.apache.spark.SparkContext
-
Cancel a given job if it's scheduled or running.
- cancelJobGroup(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Cancel active jobs for the specified group.
- cancelJobGroup(String) - Method in class org.apache.spark.SparkContext
-
Cancel active jobs for the specified group.
- cancelJobsWithTag(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Cancel active jobs that have the specified tag.
- cancelJobsWithTag(String) - Method in class org.apache.spark.SparkContext
-
Cancel active jobs that have the specified tag.
- cancelStage(int, String) - Method in class org.apache.spark.SparkContext
-
Cancel a given stage and all jobs associated with it.
- cancelStage(int) - Method in class org.apache.spark.SparkContext
-
Cancel a given stage and all jobs associated with it.
- cancelTasks(int, boolean, String) - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- canCreate(String) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
-
Check if this cluster manager instance can create scheduler components
for a certain master URL.
- canDeleteWhere(Filter[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
-
Checks whether it is possible to delete data from a data source table that matches filter
expressions.
- canDeleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDelete
-
- canDeleteWhere(Predicate[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsDeleteV2
-
Checks whether it is possible to delete data from a data source table that matches filter
expressions.
- canEqual(Object) - Static method in class org.apache.spark.ExpireDeadHosts
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.DirectPoolMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.GarbageCollectionMetrics
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.JVMHeapMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.JVMOffHeapMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.MappedPoolMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapExecutionMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapStorageMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OffHeapUnifiedMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapExecutionMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapStorageMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.OnHeapUnifiedMemory
-
- canEqual(Object) - Static method in class org.apache.spark.metrics.ProcessTreeMetrics
-
- canEqual(Object) - Static method in class org.apache.spark.ml.feature.Dot
-
- canEqual(Object) - Static method in class org.apache.spark.ml.feature.EmptyTerm
-
- canEqual(Object) - Static method in class org.apache.spark.Resubmitted
-
- canEqual(Object) - Static method in class org.apache.spark.scheduler.AllJobsCancelled
-
- canEqual(Object) - Method in class org.apache.spark.scheduler.cluster.ExecutorInfo
-
- canEqual(Object) - Static method in class org.apache.spark.scheduler.JobSucceeded
-
- canEqual(Object) - Static method in class org.apache.spark.scheduler.ResubmitFailedStages
-
- canEqual(Object) - Static method in class org.apache.spark.scheduler.StopCoordinator
-
- canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- canEqual(Object) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- canEqual(Object) - Static method in class org.apache.spark.sql.sources.AlwaysFalse
-
- canEqual(Object) - Static method in class org.apache.spark.sql.sources.AlwaysTrue
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.BinaryType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.BooleanType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.BooleanTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.ByteType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.ByteTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.CalendarIntervalType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.DateType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.DateTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.DoubleType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.DoubleTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.FloatType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.FloatTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.IntegerType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.IntegerTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.LongType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.LongTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.NullType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.ShortType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.ShortTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.StringType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.StringTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.TimestampNTZType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.TimestampType
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.TimestampTypeExpression
-
- canEqual(Object) - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
-
- canEqual(Object) - Static method in class org.apache.spark.StopMapOutputTracker
-
- canEqual(Object) - Static method in class org.apache.spark.streaming.kinesis.DefaultCredentials
-
- canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.AllReceiverIds
-
- canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- canEqual(Object) - Static method in class org.apache.spark.streaming.scheduler.StopAllReceivers
-
- canEqual(Object) - Static method in class org.apache.spark.Success
-
- canEqual(Object) - Static method in class org.apache.spark.TaskResultLost
-
- canEqual(Object) - Static method in class org.apache.spark.TaskSchedulerIsSet
-
- canEqual(Object) - Static method in class org.apache.spark.UnknownReason
-
- canEqual(Object) - Method in class org.apache.spark.util.MutablePair
-
- canHandle(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- canHandle(Driver, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcConnectionProvider
-
Checks if this connection provider instance can handle the connection initiated by the driver.
- canHandle(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Check if this dialect instance can handle a certain jdbc url.
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- canHandle(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- cannotAcquireMemoryToBuildLongHashedRelationError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotAcquireMemoryToBuildUnsafeHashedRelationError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotAddMultiPartitionsOnNonatomicPartitionTableError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotAllocateMemoryToGrowBytesToBytesMapError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotAlterTableWithAlterViewError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotAlterViewWithAlterTableError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotBroadcastTableOverMaxTableBytesError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotBroadcastTableOverMaxTableRowsError(long, long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotBuildHashedRelationLargerThan8GError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotBuildHashedRelationWithUniqueKeysExceededError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCastError(DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCastFromNullTypeError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotChangeDecimalPrecisionError(Decimal, int, int, SQLQueryContext) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- cannotChangeDecimalPrecisionError(Decimal, int, int, SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotChangeStorageLevelError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- cannotCleanReservedNamespacePropertyError(String, ParserRuleContext, String) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- cannotCleanReservedTablePropertyError(String, ParserRuleContext, String) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- cannotClearOutputDirectoryError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotClearPartitionDirectoryError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCloneOrCopyReadOnlySQLConfError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCompareCostWithTargetCostError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotConvertCatalystTypeToProtobufTypeError(Seq<String>, String, DataType, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotConvertCatalystValueToProtobufEnumTypeError(Seq<String>, String, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotConvertDataTypeToParquetTypeError(StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotConvertOrcTimestampNTZToTimestampLTZError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotConvertOrcTimestampToTimestampNTZError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotConvertProtobufTypeToCatalystTypeError(String, DataType, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotConvertProtobufTypeToSqlTypeError(String, Seq<String>, String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotConvertSqlTypeToProtobufError(String, DataType, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateArrayWithElementsExceedLimitError(long, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateColumnarReaderError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateDatabaseWithSameNameAsPreservedDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateJDBCNamespaceUsingProviderError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateJDBCTableUsingLocationError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateJDBCTableUsingProviderError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateJDBCTableWithPartitionsError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateParquetConverterForDataTypeError(DataType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateParquetConverterForDecimalTypeError(DecimalType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateParquetConverterForTypeError(DecimalType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateStagingDirError(String, IOException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotCreateTableWithBothProviderAndSerdeError(Option<String>, Option<SerdeInfo>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateTempViewUsingHiveDataSourceError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateViewNotEnoughColumnsError(TableIdentifier, Seq<String>, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotCreateViewTooManyColumnsError(TableIdentifier, Seq<String>, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotDeleteTableWhereFiltersError(Table, Predicate[]) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotDropBuiltinFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotDropDefaultDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotDropMultiPartitionsOnNonatomicPartitionTableError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotDropNonemptyDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotDropNonemptyNamespaceError(Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotEvaluateExpressionError(Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotExecuteStreamingRelationExecError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotFetchTablesOfDatabaseError(String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotFindCatalogToHandleIdentifierError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindCatalystTypeInProtobufSchemaError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindColumnError(String, String[]) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindColumnInRelationOutputError(String, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindConstructorForTypeError(String) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotFindConstructorForTypeError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotFindDescriptorFileError(String, Throwable) - Method in interface org.apache.spark.sql.errors.CompilationErrors
-
- cannotFindDescriptorFileError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindEncoderForTypeError(String) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotFindEncoderForTypeError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotFindPartitionColumnInPartitionSchemaError(StructField, StructType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotFindProtobufFieldInCatalystError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotGenerateCodeForExpressionError(Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotGenerateCodeForIncomparableTypeError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotGetEventTimeWatermarkError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotGetJdbcTypeError(DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotGetOuterPointerForInnerClassError(Class<?>) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotGetOuterPointerForInnerClassError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotGetSQLConfInSchedulerEventLoopThreadError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotHaveCircularReferencesInBeanClassError(Class<?>) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotHaveCircularReferencesInBeanClassError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotHaveCircularReferencesInClassError(String) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotHaveCircularReferencesInClassError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotInstantiateAbstractCatalogPluginClassError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotInterpolateClassIntoCodeBlockError(Object) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotLoadClassNotOnClassPathError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotLoadClassWhenRegisteringFunctionError(String, FunctionIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotLoadUserDefinedTypeError(String, String) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- cannotLoadUserDefinedTypeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotMergeClassWithOtherClassError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotMergeDecimalTypesWithIncompatibleScaleError(int, int) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- cannotMergeDecimalTypesWithIncompatibleScaleError(int, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotMergeIncompatibleDataTypesError(DataType, DataType) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- cannotMergeIncompatibleDataTypesError(DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotModifyValueOfSparkConfigError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotModifyValueOfStaticConfigError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotMutateReadOnlySQLConfError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotOperateOnHiveDataSourceFilesError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotOverwritePathBeingReadFromError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotOverwriteTableThatIsBeingReadFromError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotParseDecimalError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseIntervalError(String, Throwable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotParseJsonArraysAsStructsError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseJSONFieldError(JsonParser, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseJSONFieldError(String, String, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseStatisticAsPercentileError(String, NumberFormatException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseStringAsDataTypeError(String, String, DataType) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotParseStringAsDataTypeError(JsonParser, JsonToken, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotParseValueTypeError(String, String, SqlBaseParser.TypeConstructorContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- cannotPassTypedColumnInUntypedSelectError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotPurgeAsBreakInternalStateError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotReadFilesError(Throwable, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotReadFooterForFileError(Path, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotRecognizeHiveTypeError(ParseException, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotRefreshBuiltInFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRefreshTempFuncError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRemovePartitionDirError(Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotRemoveReservedPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotRenameTableAcrossSchemaError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRenameTableWithAlterViewError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRenameTempViewToExistingTableError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRenameTempViewWithDatabaseSpecifiedError(TableIdentifier, TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotReplaceMissingTableError(Identifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotReplaceMissingTableError(Identifier, Option<Throwable>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotResolveAttributeError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotResolveColumnGivenInputColumnsError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotResolveColumnNameAmongAttributesError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotResolveStarExpandGivenInputColumnsError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotResolveWindowReferenceError(String, SqlBaseParser.WindowClauseContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- cannotRestorePermissionsForPathError(FsPermission, Path) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotRetrieveTableOrViewNotInSameDatabaseError(Seq<QualifiedTableName>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotRunSubmitMapStageOnZeroPartitionRDDError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- cannotSafelyMergeSerdePropertiesError(Map<String, String>, Map<String, String>, Set<String>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotSaveBlockOnDecommissionedExecutorError(BlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- cannotSaveIntervalIntoExternalStorageError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotSetJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotSetTimeoutDurationError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotSetTimeoutTimestampError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotSpecifyBothJdbcTableNameAndQueryError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotSpecifyDatabaseForTempViewError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotSpecifyWindowFrameError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotTerminateGeneratorError(UnresolvedGenerator) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotTranslateExpressionToSourceFilterError(Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotTranslateNonNullValueForFieldError(int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotUnsetJDBCNamespaceWithPropertyError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotUseAllColumnsForPartitionColumnsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotUseDataTypeForPartitionColumnError(StructField) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotUseIntervalTypeInTableSchemaError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotUseInvalidJavaIdentifierAsFieldNameError(String, WalkedTypePath) - Method in interface org.apache.spark.sql.errors.ExecutionErrors
-
- cannotUseInvalidJavaIdentifierAsFieldNameError(String, WalkedTypePath) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- cannotUseMapSideCombiningWithArrayKeyError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- cannotUsePreservedDatabaseAsCurrentDatabaseError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotWriteDataToRelationsWithMultiplePathsError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotWriteNotEnoughColumnsToTableError(String, Seq<String>, Seq<Attribute>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cannotWriteTooManyColumnsToTableError(String, Seq<String>, Seq<Attribute>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- canonicalName() - Method in interface org.apache.spark.sql.connector.catalog.functions.BoundFunction
-
Returns the canonical name of this function, used to determine if functions are equivalent.
- canonicalName() - Method in class org.apache.spark.sql.connector.expressions.aggregate.UserDefinedAggregateFunc
-
- canonicalName() - Method in class org.apache.spark.sql.connector.expressions.UserDefinedScalarFunc
-
- CanonicalRandomVertexCut$() - Constructor for class org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
-
- canOnlyZipRDDsWithSamePartitionSizeError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- canOverwrite(Filter[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwrite
-
Checks whether it is possible to overwrite data from a data source table that matches filter
expressions.
- canOverwrite(Predicate[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwrite
-
- canOverwrite(Predicate[]) - Method in interface org.apache.spark.sql.connector.write.SupportsOverwriteV2
-
Checks whether it is possible to overwrite data from a data source table that matches filter
expressions.
- canRenameConflictingMetadataColumns() - Method in interface org.apache.spark.sql.connector.catalog.SupportsMetadataColumns
-
Determines how this data source handles name conflicts between metadata and data columns.
- canUpCast(DataType, DataType) - Static method in class org.apache.spark.sql.types.UpCastRule
-
Returns true iff we can safely up-cast the from
type to to
type without any truncating or
precision lose or possible runtime failures.
- capabilities() - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- capabilities() - Method in interface org.apache.spark.sql.connector.catalog.Table
-
Returns the set of capabilities for this table.
- capabilities() - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
-
- cardinality(Column) - Static method in class org.apache.spark.sql.functions
-
Returns length of array or map.
- cardinality() - Method in class org.apache.spark.util.sketch.BloomFilter
-
- cartesian(JavaRDDLike<U, ?>) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of
elements (a, b) where a is in this
and b is in other
.
- cartesian(RDD<U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
-
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of
elements (a, b) where a is in this
and b is in other
.
- CaseInsensitiveStringMap - Class in org.apache.spark.sql.util
-
Case-insensitive map of string keys to string values.
- CaseInsensitiveStringMap(Map<String, String>) - Constructor for class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- caseSensitive() - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
Whether to do a case sensitive comparison over the stop words.
- cast(DataType) - Method in class org.apache.spark.sql.Column
-
Casts the column to a different data type.
- cast(String) - Method in class org.apache.spark.sql.Column
-
Casts the column to a different data type, using the canonical string representation
of the type.
- Cast - Class in org.apache.spark.sql.connector.expressions
-
Represents a cast expression in the public logical expression API.
- Cast(Expression, DataType) - Constructor for class org.apache.spark.sql.connector.expressions.Cast
-
- castingCauseOverflowError(String, DataType, DataType) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- castingCauseOverflowError(Object, DataType, DataType) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- castingCauseOverflowErrorInTableInsert(DataType, DataType, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- castPartitionSpec(String, DataType, SQLConf) - Static method in class org.apache.spark.sql.util.PartitioningUtils
-
- Catalog - Class in org.apache.spark.sql.catalog
-
Catalog interface for Spark.
- Catalog() - Constructor for class org.apache.spark.sql.catalog.Catalog
-
- catalog() - Method in class org.apache.spark.sql.catalog.Database
-
- catalog() - Method in class org.apache.spark.sql.catalog.Function
-
- catalog() - Method in class org.apache.spark.sql.catalog.Table
-
- catalog() - Method in class org.apache.spark.sql.SparkSession
-
- CatalogAndIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- CatalogAndIdentifier() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier
-
- CatalogAndIdentifier$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndIdentifier$
-
- CatalogAndMultipartIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- CatalogAndNamespace() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- CatalogAndNamespace() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace
-
- CatalogAndNamespace$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.CatalogAndNamespace$
-
- CatalogExtension - Interface in org.apache.spark.sql.connector.catalog
-
An API to extend the Spark built-in session catalog.
- catalogFailToCallPublicNoArgConstructorError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- catalogFailToFindPublicNoArgConstructorError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- CatalogHelper(CatalogPlugin) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.CatalogHelper
-
- catalogManager() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- CatalogMetadata - Class in org.apache.spark.sql.catalog
-
A catalog in Spark, as returned by the
listCatalogs
method defined in
Catalog
.
- CatalogMetadata(String, String) - Constructor for class org.apache.spark.sql.catalog.CatalogMetadata
-
- CatalogNotFoundException - Exception in org.apache.spark.sql.connector.catalog
-
- CatalogNotFoundException(String, Throwable) - Constructor for exception org.apache.spark.sql.connector.catalog.CatalogNotFoundException
-
- CatalogNotFoundException(String) - Constructor for exception org.apache.spark.sql.connector.catalog.CatalogNotFoundException
-
- catalogOperationNotSupported(CatalogPlugin, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- CatalogPlugin - Interface in org.apache.spark.sql.connector.catalog
-
A marker interface to provide a catalog implementation for Spark.
- catalogPluginClassNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- catalogPluginClassNotFoundForCatalogError(String, String, Exception) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- catalogPluginClassNotImplementedError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- Catalogs - Class in org.apache.spark.sql.connector.catalog
-
- Catalogs() - Constructor for class org.apache.spark.sql.connector.catalog.Catalogs
-
- catalogString() - Method in class org.apache.spark.sql.types.ArrayType
-
- catalogString() - Static method in class org.apache.spark.sql.types.BinaryType
-
- catalogString() - Static method in class org.apache.spark.sql.types.BooleanType
-
- catalogString() - Static method in class org.apache.spark.sql.types.ByteType
-
- catalogString() - Static method in class org.apache.spark.sql.types.CalendarIntervalType
-
- catalogString() - Method in class org.apache.spark.sql.types.DataType
-
String representation for the type saved in external catalogs.
- catalogString() - Static method in class org.apache.spark.sql.types.DateType
-
- catalogString() - Static method in class org.apache.spark.sql.types.DoubleType
-
- catalogString() - Static method in class org.apache.spark.sql.types.FloatType
-
- catalogString() - Static method in class org.apache.spark.sql.types.IntegerType
-
- catalogString() - Static method in class org.apache.spark.sql.types.LongType
-
- catalogString() - Method in class org.apache.spark.sql.types.MapType
-
- catalogString() - Static method in class org.apache.spark.sql.types.NullType
-
- catalogString() - Static method in class org.apache.spark.sql.types.ShortType
-
- catalogString() - Static method in class org.apache.spark.sql.types.StringType
-
- catalogString() - Method in class org.apache.spark.sql.types.StructType
-
- catalogString() - Static method in class org.apache.spark.sql.types.TimestampNTZType
-
- catalogString() - Static method in class org.apache.spark.sql.types.TimestampType
-
- catalogString() - Method in class org.apache.spark.sql.types.UserDefinedType
-
- CatalogV2Implicits - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits() - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits
-
- CatalogV2Implicits.BucketSpecHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.CatalogHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.ColumnsHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.FunctionIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.IdentifierHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.MultipartIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.NamespaceHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.PartitionTypeHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.TableIdentifierHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Implicits.TransformHelper - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Util - Class in org.apache.spark.sql.connector.catalog
-
- CatalogV2Util() - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
- CatalystScan - Interface in org.apache.spark.sql.sources
-
::Experimental::
An interface for experimenting with a more direct connection to the query planner.
- Categorical() - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
-
- categoricalCols() - Method in class org.apache.spark.ml.feature.FeatureHasher
-
Numeric columns to treat as categorical features.
- categoricalFeaturesInfo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- CategoricalSplit - Class in org.apache.spark.ml.tree
-
Split which tests a categorical feature.
- categories() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- categories() - Method in class org.apache.spark.mllib.tree.model.Split
-
- categoryMaps() - Method in class org.apache.spark.ml.feature.VectorIndexerModel
-
- categorySizes() - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- cause() - Method in exception org.apache.spark.sql.AnalysisException
-
- cause() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
-
- CausedBy - Class in org.apache.spark.util
-
Extractor Object for pulling out the root cause of an error.
- CausedBy() - Constructor for class org.apache.spark.util.CausedBy
-
- cbrt(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the cube-root of the given value.
- cbrt(String) - Static method in class org.apache.spark.sql.functions
-
Computes the cube-root of the given column.
- ceil(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Computes the ceiling of the given value of e
to scale
decimal places.
- ceil(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the ceiling of the given value of e
to 0 decimal places.
- ceil(String) - Static method in class org.apache.spark.sql.functions
-
Computes the ceiling of the given value of e
to 0 decimal places.
- ceil() - Method in class org.apache.spark.sql.types.Decimal
-
- ceiling(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Computes the ceiling of the given value of e
to scale
decimal places.
- ceiling(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the ceiling of the given value of e
to 0 decimal places.
- censorCol() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- censorCol() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- censorCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
-
Param for censor column name.
- centerMatrix() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
-
- chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, T, T>>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<U>>, Function0<Parsers.Parser<Function2<T, U, T>>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- chainr1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, U, U>>>, Function2<T, U, U>, U) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- changePrecision(int, int) - Method in class org.apache.spark.sql.types.Decimal
-
Update precision and scale while keeping our value the same, and return true if successful.
- channel() - Method in interface org.apache.spark.shuffle.api.WritableByteChannelWrapper
-
The underlying channel to write bytes into.
- channelRead0(ChannelHandlerContext, byte[]) - Method in class org.apache.spark.api.r.RBackendAuthHandler
-
- char_length(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the character length of string data or number of bytes of binary data.
- character_length(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the character length of string data or number of bytes of binary data.
- charOrVarcharTypeAsStringUnsupportedError() - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- charOrVarcharTypeAsStringUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- CharType - Class in org.apache.spark.sql.types
-
- CharType(int) - Constructor for class org.apache.spark.sql.types.CharType
-
- charTypeMissingLengthError(String, SqlBaseParser.PrimitiveDataTypeContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- checkAndGetK8sMasterUrl(String) - Static method in class org.apache.spark.util.Utils
-
Check the validity of the given Kubernetes master URL and return the resolved URL.
- checkColumnNameDuplication(Seq<String>, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Checks if input column names have duplicate identifiers.
- checkColumnNameDuplication(Seq<String>, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Checks if input column names have duplicate identifiers.
- checkColumnType(StructType, String, DataType, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of the required data type.
- checkColumnTypes(StructType, String, Seq<DataType>, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of one of the require data types.
- checkDataColumns(RFormula, Dataset<?>) - Static method in class org.apache.spark.ml.r.RWrapperUtils
-
DataFrame column check.
- checkFileExists(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
Check if the file exists at the given path.
- checkHost(String) - Static method in class org.apache.spark.util.Utils
-
Checks if the host contains only valid hostname/ip without port
NOTE: Incase of IPV6 ip it should be enclosed inside []
- checkHostPort(String) - Static method in class org.apache.spark.util.Utils
-
- checkIntegers(Dataset<?>, String) - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
-
Attempts to safely cast a user/item id to an Int.
- checkNumericType(StructType, String, String) - Static method in class org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of the numeric data type.
- checkOffHeapEnabled(SparkConf, long) - Static method in class org.apache.spark.util.Utils
-
return 0 if MEMORY_OFFHEAP_ENABLED is false.
- checkpoint() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Mark this RDD for checkpointing.
- checkpoint() - Method in class org.apache.spark.graphx.Graph
-
Mark this Graph for checkpointing.
- checkpoint() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
- checkpoint() - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- checkpoint() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
- checkpoint() - Method in class org.apache.spark.rdd.HadoopRDD
-
- checkpoint() - Method in class org.apache.spark.rdd.RDD
-
Mark this RDD for checkpointing.
- checkpoint() - Method in class org.apache.spark.sql.Dataset
-
Eagerly checkpoint a Dataset and return the new Dataset.
- checkpoint(boolean) - Method in class org.apache.spark.sql.Dataset
-
Returns a checkpointed version of this Dataset.
- checkpoint(Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Enable periodic checkpointing of RDDs of this DStream.
- checkpoint(String) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Sets the context to periodically checkpoint the DStream operations for master
fault-tolerance.
- checkpoint(Duration) - Method in class org.apache.spark.streaming.dstream.DStream
-
Enable periodic checkpointing of RDDs of this DStream
- checkpoint(String) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Set the context to periodically checkpoint the DStream operations for driver
fault-tolerance.
- checkpointCleaned(long) - Method in interface org.apache.spark.CleanerListener
-
- checkpointDirectoryHasNotBeenSetInSparkContextError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- Checkpointed() - Static method in class org.apache.spark.rdd.CheckpointState
-
- checkpointFailedToSaveError(int, Path) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- CheckpointingInProgress() - Static method in class org.apache.spark.rdd.CheckpointState
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- checkpointInterval() - Method in class org.apache.spark.ml.clustering.LDA
-
- checkpointInterval() - Method in class org.apache.spark.ml.clustering.LDAModel
-
- checkpointInterval() - Method in interface org.apache.spark.ml.param.shared.HasCheckpointInterval
-
Param for set checkpoint interval (>= 1) or disable checkpoint (-1).
- checkpointInterval() - Method in class org.apache.spark.ml.recommendation.ALS
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
-
- checkpointInterval() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- checkpointInterval() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- checkpointLocationNotSpecifiedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- checkpointRDDBlockIdNotFoundError(RDDBlockId) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- checkpointRDDHasDifferentNumberOfPartitionsFromOriginalRDDError(int, int, int, int) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- CheckpointReader - Class in org.apache.spark.streaming
-
- CheckpointReader() - Constructor for class org.apache.spark.streaming.CheckpointReader
-
- CheckpointState - Class in org.apache.spark.rdd
-
Enumeration to manage state transitions of an RDD through checkpointing
- CheckpointState() - Constructor for class org.apache.spark.rdd.CheckpointState
-
- checkSchemaColumnNameDuplication(DataType, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Checks if an input schema has duplicate column names.
- checkSchemaColumnNameDuplication(StructType, Function2<String, String, Object>) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Checks if an input schema has duplicate column names.
- checkSingleVsMultiColumnParams(Params, Seq<Param<?>>, Seq<Param<?>>) - Static method in class org.apache.spark.ml.param.ParamValidators
-
Utility for Param validity checks for Transformers which have both single- and multi-column
support.
- checkSpeculatableTasks(long) - Method in interface org.apache.spark.scheduler.Schedulable
-
- checkState(boolean, Function0<String>) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
- checkThresholdConsistency() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
If threshold
and thresholds
are both set, ensures they are consistent.
- checkTransformDuplication(Seq<Transform>, String, boolean) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Checks if the partitioning transforms are being duplicated or not.
- checkUIViewPermissions() - Method in interface org.apache.spark.status.api.v1.BaseAppResource
-
- checkUIViewPermissions(String, Option<String>, String) - Method in interface org.apache.spark.status.api.v1.UIRoot
-
- child() - Method in class org.apache.spark.sql.connector.expressions.filter.Not
-
- child() - Method in class org.apache.spark.sql.sources.Not
-
- CHILD_CLUSTERS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- CHILD_CONNECTION_TIMEOUT - Static variable in class org.apache.spark.launcher.SparkLauncher
-
Maximum time (in ms) to wait for a child process to connect back to the launcher server
when using @link{#start()}.
- CHILD_NODES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- CHILD_PROCESS_LOGGER_NAME - Static variable in class org.apache.spark.launcher.SparkLauncher
-
Logger name to use when launching a child process.
- ChildFirstURLClassLoader - Class in org.apache.spark.util
-
A mutable class loader that gives preference to its own URLs over the parent class loader
when loading classes and resources.
- ChildFirstURLClassLoader(URL[], ClassLoader) - Constructor for class org.apache.spark.util.ChildFirstURLClassLoader
-
- ChildFirstURLClassLoader(URL[], ClassLoader, ClassLoader) - Constructor for class org.apache.spark.util.ChildFirstURLClassLoader
-
Specify the grandparent if there is a need to load in the order of
`grandparent -> urls (child) -> parent`.
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Avg
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Count
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.CountStar
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.GeneralAggregateFunc
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Max
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Min
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Sum
-
- children() - Method in class org.apache.spark.sql.connector.expressions.aggregate.UserDefinedAggregateFunc
-
- children() - Method in class org.apache.spark.sql.connector.expressions.Cast
-
- children() - Method in interface org.apache.spark.sql.connector.expressions.Expression
-
Returns an array of the children of this node.
- children() - Method in class org.apache.spark.sql.connector.expressions.Extract
-
- children() - Method in class org.apache.spark.sql.connector.expressions.GeneralScalarExpression
-
- children() - Method in interface org.apache.spark.sql.connector.expressions.Literal
-
- children() - Method in interface org.apache.spark.sql.connector.expressions.NamedReference
-
- children() - Method in interface org.apache.spark.sql.connector.expressions.SortOrder
-
- children() - Method in interface org.apache.spark.sql.connector.expressions.Transform
-
- children() - Method in class org.apache.spark.sql.connector.expressions.UserDefinedScalarFunc
-
- chiSqFunc() - Method in class org.apache.spark.mllib.stat.test.ChiSqTest.Method
-
- ChiSqSelector - Class in org.apache.spark.ml.feature
-
- ChiSqSelector(String) - Constructor for class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- ChiSqSelector() - Constructor for class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- ChiSqSelector - Class in org.apache.spark.mllib.feature
-
Creates a ChiSquared feature selector.
- ChiSqSelector() - Constructor for class org.apache.spark.mllib.feature.ChiSqSelector
-
- ChiSqSelector(int) - Constructor for class org.apache.spark.mllib.feature.ChiSqSelector
-
The is the same to call this() and setNumTopFeatures(numTopFeatures)
- ChiSqSelectorModel - Class in org.apache.spark.ml.feature
-
- ChiSqSelectorModel - Class in org.apache.spark.mllib.feature
-
Chi Squared selector model.
- ChiSqSelectorModel(int[]) - Constructor for class org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- ChiSqSelectorModel.ChiSqSelectorModelWriter - Class in org.apache.spark.ml.feature
-
- ChiSqSelectorModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.feature
-
- ChiSqSelectorModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.feature
-
Model data for import/export
- ChiSqSelectorModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.feature
-
- ChiSqSelectorModelWriter(ChiSqSelectorModel) - Constructor for class org.apache.spark.ml.feature.ChiSqSelectorModel.ChiSqSelectorModelWriter
-
- chiSqTest(Vector, Vector) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's chi-squared goodness of fit test of the observed data against the
expected distribution.
- chiSqTest(Vector) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's chi-squared goodness of fit test of the observed data against the uniform
distribution, with each category having an expected frequency of 1 / observed.size
.
- chiSqTest(Matrix) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's independence test on the input contingency matrix, which cannot contain
negative entries or columns or rows that sum up to 0.
- chiSqTest(RDD<LabeledPoint>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's independence test for every feature against the label across the input RDD.
- chiSqTest(JavaRDD<LabeledPoint>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of chiSqTest()
- ChiSqTest - Class in org.apache.spark.mllib.stat.test
-
Conduct the chi-squared test for the input RDDs using the specified method.
- ChiSqTest() - Constructor for class org.apache.spark.mllib.stat.test.ChiSqTest
-
- ChiSqTest.Method - Class in org.apache.spark.mllib.stat.test
-
param: name String name for the method.
- ChiSqTest.Method$ - Class in org.apache.spark.mllib.stat.test
-
- ChiSqTest.NullHypothesis$ - Class in org.apache.spark.mllib.stat.test
-
- ChiSqTestResult - Class in org.apache.spark.mllib.stat.test
-
Object containing the test results for the chi-squared hypothesis test.
- chiSquared(Vector, Vector, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
-
- chiSquaredFeatures(RDD<LabeledPoint>, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
-
Conduct Pearson's independence test for each feature against the label across the input RDD.
- chiSquaredMatrix(Matrix, String) - Static method in class org.apache.spark.mllib.stat.test.ChiSqTest
-
- ChiSquareTest - Class in org.apache.spark.ml.stat
-
Chi-square hypothesis testing for categorical data.
- ChiSquareTest() - Constructor for class org.apache.spark.ml.stat.ChiSquareTest
-
- chmod700(File) - Static method in class org.apache.spark.util.Utils
-
JDK equivalent of chmod 700 file
.
- CholeskyDecomposition - Class in org.apache.spark.mllib.linalg
-
Compute Cholesky decomposition.
- CholeskyDecomposition() - Constructor for class org.apache.spark.mllib.linalg.CholeskyDecomposition
-
- chr(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the ASCII character having the binary equivalent to n
.
- chunkId() - Method in class org.apache.spark.storage.ShuffleBlockChunkId
-
- cipherStream() - Method in interface org.apache.spark.security.CryptoStreamUtils.BaseErrorHandler
-
The encrypted stream that may get into an unhealthy state.
- classDoesNotImplementUserDefinedAggregateFunctionError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- classForName(String, boolean, boolean) - Method in interface org.apache.spark.util.SparkClassUtils
-
Preferred alternative to Class.forName(className), as well as
Class.forName(className, initialize, loader) with current thread's ContextClassLoader.
- classForName(String, boolean, boolean) - Static method in class org.apache.spark.util.Utils
-
- classForName$default$2() - Static method in class org.apache.spark.util.Utils
-
- classForName$default$3() - Static method in class org.apache.spark.util.Utils
-
- classHasUnexpectedSerializerError(String, Expression) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- Classification() - Static method in class org.apache.spark.mllib.tree.configuration.Algo
-
- ClassificationLoss - Interface in org.apache.spark.mllib.tree.loss
-
- ClassificationModel<FeaturesType,M extends ClassificationModel<FeaturesType,M>> - Class in org.apache.spark.ml.classification
-
- ClassificationModel() - Constructor for class org.apache.spark.ml.classification.ClassificationModel
-
- ClassificationModel - Interface in org.apache.spark.mllib.classification
-
Represents a classification model that predicts to which of a set of categories an example
belongs.
- ClassificationSummary - Interface in org.apache.spark.ml.classification
-
Abstraction for multiclass classification results for a given model.
- Classifier<FeaturesType,E extends Classifier<FeaturesType,E,M>,M extends ClassificationModel<FeaturesType,M>> - Class in org.apache.spark.ml.classification
-
Single-label binary or multiclass classification.
- Classifier() - Constructor for class org.apache.spark.ml.classification.Classifier
-
- classifier() - Method in class org.apache.spark.ml.classification.OneVsRest
-
- classifier() - Method in class org.apache.spark.ml.classification.OneVsRestModel
-
- classifier() - Method in interface org.apache.spark.ml.classification.OneVsRestParams
-
param for the base binary classifier that we reduce multiclass classification into.
- ClassifierParams - Interface in org.apache.spark.ml.classification
-
(private[spark]) Params for classification.
- ClassifierTypeTrait - Interface in org.apache.spark.ml.classification
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- classifyException(String, Throwable) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Gets a dialect exception, classifies it and wraps it by AnalysisException
.
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- classifyException(String, Throwable) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- classIsLoadable(String) - Method in interface org.apache.spark.util.SparkClassUtils
-
Determines whether the provided class is loadable in the current thread.
- classIsLoadable(String) - Static method in class org.apache.spark.util.Utils
-
- className() - Method in class org.apache.spark.ExceptionFailure
-
- className() - Static method in class org.apache.spark.ml.linalg.JsonMatrixConverter
-
Unique class name for identifying JSON object encoded by this class.
- className() - Method in class org.apache.spark.sql.catalog.Function
-
- CLASSPATH_ENTRIES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- classpathEntries() - Method in class org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- classTag() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
- classTag() - Method in class org.apache.spark.api.java.JavaPairRDD
-
- classTag() - Method in class org.apache.spark.api.java.JavaRDD
-
- classTag() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
- classTag() - Method in class org.apache.spark.sql.Dataset
-
- classTag() - Method in class org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- classTag() - Method in interface org.apache.spark.storage.memory.MemoryEntry
-
- classTag() - Method in class org.apache.spark.storage.memory.SerializedMemoryEntry
-
- classTag() - Method in class org.apache.spark.streaming.api.java.JavaDStream
-
- classTag() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
- classTag() - Method in class org.apache.spark.streaming.api.java.JavaInputDStream
-
- classTag() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
- classTag() - Method in class org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- classUnsupportedByMapObjectsError(Class<?>) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- classWithoutPublicNonArgumentConstructorError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- clean(long, boolean) - Method in class org.apache.spark.streaming.util.WriteAheadLog
-
Clean all the records that are older than the threshold time.
- clean(Object, boolean, boolean) - Static method in class org.apache.spark.util.ClosureCleaner
-
Clean the given closure in place.
- CleanAccum - Class in org.apache.spark
-
- CleanAccum(long) - Constructor for class org.apache.spark.CleanAccum
-
- CleanBroadcast - Class in org.apache.spark
-
- CleanBroadcast(long) - Constructor for class org.apache.spark.CleanBroadcast
-
- CleanCheckpoint - Class in org.apache.spark
-
- CleanCheckpoint(int) - Constructor for class org.apache.spark.CleanCheckpoint
-
- CleanerListener - Interface in org.apache.spark
-
Listener class used when any item has been cleaned by the Cleaner class.
- cleaning() - Method in class org.apache.spark.status.LiveStage
-
- CleanRDD - Class in org.apache.spark
-
- CleanRDD(int) - Constructor for class org.apache.spark.CleanRDD
-
- CleanShuffle - Class in org.apache.spark
-
- CleanShuffle(int) - Constructor for class org.apache.spark.CleanShuffle
-
- cleanShuffleDependencies(boolean) - Method in class org.apache.spark.rdd.RDD
-
Removes an RDD's shuffles and it's non-persisted ancestors.
- CleanSparkListener - Class in org.apache.spark
-
- CleanSparkListener(SparkListener) - Constructor for class org.apache.spark.CleanSparkListener
-
- cleanupApplication() - Method in interface org.apache.spark.shuffle.api.ShuffleDriverComponents
-
Called once at the end of the Spark application to clean up any existing shuffle state.
- cleanupOldBlocks(long) - Method in interface org.apache.spark.streaming.receiver.ReceivedBlockHandler
-
Cleanup old blocks older than the given threshold time
- cleanUpSourceFilesUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- CleanupTask - Interface in org.apache.spark
-
Classes that represent cleaning tasks.
- CleanupTaskWeakReference - Class in org.apache.spark
-
A WeakReference associated with a CleanupTask.
- CleanupTaskWeakReference(CleanupTask, Object, ReferenceQueue<Object>) - Constructor for class org.apache.spark.CleanupTaskWeakReference
-
- clear(Param<?>) - Method in interface org.apache.spark.ml.param.Params
-
Clears the user-supplied value for the input param.
- clear() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- clear() - Method in class org.apache.spark.sql.util.ExecutionListenerManager
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- clear() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- clear() - Static method in class org.apache.spark.util.AccumulatorContext
-
- clearAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
int64 accumulator_id = 2;
- clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- clearAccumulatorUpdates() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- clearActive() - Static method in class org.apache.spark.sql.SQLContext
-
- clearActiveSession() - Static method in class org.apache.spark.sql.SparkSession
-
Clears the active SparkSession for current thread.
- clearActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 active_tasks = 9;
- clearAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional string address = 1;
- clearAddresses() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- clearAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 add_time = 20;
- clearAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int64 add_time = 5;
- clearAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_removals_time_ms = 6;
- clearAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_updates_time_ms = 4;
- clearAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
int64 amount = 2;
- clearAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
double amount = 2;
- clearAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string app_spark_version = 8;
- clearAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 attempt = 3;
- clearAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 attempt = 3;
- clearAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string attempt_id = 1;
- clearAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 attempt_id = 3;
- clearAttempts() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- clearAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clearBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool barrier = 4;
- clearBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_duration = 6;
- clearBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_id = 5;
- clearBlacklistedInStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- clearBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string block_name = 1;
- clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_read = 18;
- clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- clearBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 bytes_read = 1;
- clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_written = 20;
- clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 bytes_written = 1;
- clearBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 bytes_written = 1;
- clearCache() - Method in class org.apache.spark.sql.catalog.Catalog
-
Removes all cached tables from the in-memory cache.
- clearCache() - Method in class org.apache.spark.sql.SQLContext
-
Removes all cached tables from the in-memory cache.
- clearCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool cached = 3;
- clearCallSite() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Pass-through to SparkContext.setCallSite.
- clearCallSite() - Method in class org.apache.spark.SparkContext
-
Clear the thread-local property for overriding the call sites
of actions and RDDs.
- clearCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string callsite = 5;
- clearChildClusters() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- clearChildNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- clearClasspathEntries() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- clearCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- clearCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 commit_time_ms = 7;
- clearCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
bool completed = 7;
- clearCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 completed_tasks = 11;
- clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 completion_time = 5;
- clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional int64 completion_time = 9;
- clearCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 completion_time = 12;
- clearCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_granted = 3;
- clearCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_per_executor = 5;
- clearCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- clearCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 corrupt_merged_block_chunks = 1;
- clearCurrentDefaultValue() - Method in class org.apache.spark.sql.types.StructField
-
Clears the StructField of its current default value, if any.
- clearCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- clearDataDistribution() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- clearDefaultSession() - Static method in class org.apache.spark.sql.SparkSession
-
Clears the default SparkSession that is returned by the builder.
- clearDependencies() - Method in class org.apache.spark.rdd.CoGroupedRDD
-
- clearDependencies() - Method in class org.apache.spark.rdd.ShuffledRDD
-
- clearDependencies() - Method in class org.apache.spark.rdd.UnionRDD
-
- clearDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string desc = 3;
- clearDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string desc = 3;
- clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string description = 3;
- clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
optional string description = 1;
- clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string description = 1;
- clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string description = 3;
- clearDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string description = 40;
- clearDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool deserialized = 7;
- clearDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string details = 4;
- clearDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string details = 41;
- clearDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string discoveryScript = 3;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double disk_bytes_spilled = 17;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 disk_bytes_spilled = 14;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 disk_bytes_spilled = 22;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 disk_bytes_spilled = 24;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- clearDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 disk_bytes_spilled = 9;
- clearDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 disk_size = 9;
- clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 disk_used = 6;
- clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 disk_used = 4;
- clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 disk_used = 4;
- clearDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 disk_used = 7;
- clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 duration = 5;
- clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double duration = 5;
- clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 duration = 7;
- clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 duration = 7;
- clearDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- clearDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clearEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- clearEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- clearEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string end_offset = 3;
- clearEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 end_time = 3;
- clearEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional int64 end_timestamp = 7;
- clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string error_message = 10;
- clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string error_message = 14;
- clearErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string error_message = 14;
- clearEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clearException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string exception = 5;
- clearExcludedInStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- clearExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
int64 execution_id = 1;
- clearExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 execution_id = 1;
- clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_cpu_time = 9;
- clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_cpu_time = 17;
- clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_cpu_time = 19;
- clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- clearExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_cpu_time = 4;
- clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_cpu_time = 7;
- clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_cpu_time = 15;
- clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_cpu_time = 17;
- clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- clearExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_cpu_time = 2;
- clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_time = 6;
- clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_time = 14;
- clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_time = 16;
- clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- clearExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_time = 1;
- clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
optional string executor_id = 3;
- clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string executor_id = 2;
- clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string executor_id = 8;
- clearExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string executor_id = 8;
- clearExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clearExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- clearExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- clearExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- clearExecutorResourceRequests() - Method in class org.apache.spark.resource.ResourceProfileBuilder
-
- clearExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_run_time = 8;
- clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_run_time = 16;
- clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_run_time = 18;
- clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- clearExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_run_time = 3;
- clearExecutors() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- clearExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 failed_tasks = 2;
- clearFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 failed_tasks = 10;
- clearFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string failure_reason = 13;
- clearFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- clearFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 fetch_wait_time = 3;
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- clearField(Descriptors.FieldDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- clearFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 first_task_launched_time = 11;
- clearFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 from_id = 1;
- clearFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 from_id = 1;
- clearFunctions() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double getting_result_time = 13;
- clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 getting_result_time = 18;
- clearGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- clearHadoopProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- clearHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool has_metrics = 15;
- clearHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string host = 9;
- clearHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string host = 9;
- clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string host_port = 2;
- clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string host_port = 2;
- clearHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string host_port = 3;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
int64 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
int32 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
int32 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
int64 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
int64 id = 1;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string id = 2;
- clearId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string id = 1;
- clearIncomingEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- clearIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 index = 2;
- clearIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 index = 2;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- clearInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_bytes = 5;
- clearInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_bytes = 24;
- clearInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_bytes_read = 26;
- clearInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- clearInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_records = 6;
- clearInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_records = 25;
- clearInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_records_read = 27;
- clearInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double input_rows_per_second = 6;
- clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_active = 3;
- clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
bool is_active = 3;
- clearIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
bool is_active = 4;
- clearIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_blacklisted = 18;
- clearIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_blacklisted_for_stage = 15;
- clearIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_excluded = 30;
- clearIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_excluded_for_stage = 17;
- clearIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
bool is_shuffle_push_enabled = 63;
- clearJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_home = 2;
- clearJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_version = 1;
- clearJobGroup() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Clear the current thread's job group ID and its description.
- clearJobGroup() - Method in class org.apache.spark.SparkContext
-
Clear the current thread's job group ID and its description.
- clearJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string job_group = 7;
- clearJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
All IDs are int64 for extendability, even when they are currently int32 in Spark.
- clearJobIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- clearJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clearJobTags() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Clear the current thread's job tags.
- clearJobTags() - Method in class org.apache.spark.SparkContext
-
Clear the current thread's job tags.
- clearJobTags() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double jvm_gc_time = 11;
- clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 jvm_gc_time = 19;
- clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 jvm_gc_time = 21;
- clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- clearJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 jvm_gc_time = 6;
- clearKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- clearKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 killed_tasks = 4;
- clearKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clearKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- clearLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 last_updated = 4;
- clearLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string latest_offset = 4;
- clearLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 launch_time = 5;
- clearLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 launch_time = 5;
- clearLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- clearLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_blocks_fetched = 2;
- clearLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_bytes_read = 6;
- clearLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- clearLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- clearLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_blocks_fetched = 4;
- clearLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- clearLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_bytes_read = 8;
- clearLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- clearLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_chunks_fetched = 6;
- clearMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 max_cores = 4;
- clearMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 max_memory = 19;
- clearMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 max_tasks = 8;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double memory_bytes_spilled = 16;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 memory_bytes_spilled = 13;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 memory_bytes_spilled = 21;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 memory_bytes_spilled = 23;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- clearMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 memory_bytes_spilled = 8;
- clearMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- clearMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 memory_per_executor_mb = 6;
- clearMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_remaining = 3;
- clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 memory_used = 5;
- clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_used = 2;
- clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 memory_used = 3;
- clearMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 memory_used = 6;
- clearMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 memory_used_bytes = 8;
- clearMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 mem_size = 8;
- clearMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- clearMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 merged_fetch_fallback_count = 2;
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- clearMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- clearMetricsProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- clearMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string metric_type = 3;
- clearMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clearMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
bool metric_values_is_null = 13;
- clearModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
optional string name = 1;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
optional string name = 1;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string name = 2;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string name = 1;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string name = 39;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string name = 1;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string name = 1;
- clearName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string name = 3;
- clearNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- clearNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- clearNodes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- clearNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_stages = 16;
- clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_tasks = 10;
- clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_active_tasks = 2;
- clearNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_active_tasks = 5;
- clearNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_cached_partitions = 4;
- clearNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_indices = 15;
- clearNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_completed_indices = 9;
- clearNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_jobs = 1;
- clearNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_stages = 2;
- clearNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_stages = 17;
- clearNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_tasks = 11;
- clearNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_completed_tasks = 3;
- clearNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_complete_tasks = 6;
- clearNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_stages = 19;
- clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_tasks = 13;
- clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_failed_tasks = 4;
- clearNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_failed_tasks = 7;
- clearNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
int64 num_input_rows = 5;
- clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_killed_tasks = 14;
- clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_killed_tasks = 5;
- clearNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_killed_tasks = 8;
- clearNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
int64 num_output_rows = 2;
- clearNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_partitions = 3;
- clearNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_dropped_by_watermark = 9;
- clearNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_removed = 5;
- clearNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_total = 2;
- clearNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_updated = 3;
- clearNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_shuffle_partitions = 10;
- clearNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_stages = 18;
- clearNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_tasks = 12;
- clearNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_state_store_instances = 11;
- clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_tasks = 9;
- clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_tasks = 1;
- clearNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_tasks = 4;
- clearObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clearOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_remaining = 8;
- clearOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_used = 6;
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- clearOneof(Descriptors.OneofDescriptor) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- clearOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_remaining = 7;
- clearOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_used = 5;
- clearOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
optional string operator_name = 1;
- clearOutgoingEdges() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_bytes = 7;
- clearOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_bytes = 26;
- clearOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_bytes_written = 28;
- clearOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- clearOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- clearOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_records = 8;
- clearOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_records = 27;
- clearOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_records_written = 29;
- clearPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 partition_id = 4;
- clearPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 partition_id = 4;
- clearPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double peak_execution_memory = 15;
- clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 peak_execution_memory = 23;
- clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 peak_execution_memory = 25;
- clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- clearPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 peak_execution_memory = 10;
- clearPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- clearPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- clearPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string physical_plan_description = 5;
- clearProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double processed_rows_per_second = 7;
- clearProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- clearProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- clearQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
optional string quantile = 3;
- clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- clearQuantiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- clearRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 rdd_blocks = 4;
- clearRddIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- clearReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- clearReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_read = 19;
- clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 records_read = 2;
- clearRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 records_read = 7;
- clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_written = 21;
- clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 records_written = 2;
- clearRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 records_written = 3;
- clearRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- clearRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_blocks_fetched = 1;
- clearRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- clearRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read = 4;
- clearRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- clearRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read_to_disk = 5;
- clearRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- clearRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_blocks_fetched = 3;
- clearRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- clearRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_bytes_read = 7;
- clearRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- clearRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_chunks_fetched = 5;
- clearRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- clearRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_reqs_duration = 9;
- clearRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- clearRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_reqs_duration = 8;
- clearRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string remove_reason = 22;
- clearRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional int64 remove_time = 21;
- clearRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional int64 remove_time = 6;
- clearResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string resource_name = 1;
- clearResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
optional string resource_name = 1;
- clearResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 resource_profile_id = 29;
- clearResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 resource_profile_id = 49;
- clearResourceProfiles() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- clearResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clearResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 result_fetch_start = 6;
- clearResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_fetch_start = 6;
- clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_serialization_time = 12;
- clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_serialization_time = 20;
- clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_serialization_time = 22;
- clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- clearResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_serialization_time = 7;
- clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_size = 10;
- clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_size = 18;
- clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_size = 20;
- clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- clearResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_size = 5;
- clearRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- clearRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 root_execution_id = 2;
- clearRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- clearRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string run_id = 3;
- clearRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string run_id = 2;
- clearRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- clearScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string scala_version = 3;
- clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double scheduler_delay = 14;
- clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 scheduler_delay = 17;
- clearSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- clearSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string scheduling_pool = 42;
- clearShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_bytes_written = 37;
- clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_corrupt_merged_block_chunks = 33;
- clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 53;
- clearShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 42;
- clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_fetch_wait_time = 26;
- clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_fetch_wait_time = 30;
- clearShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_fetch_wait_time = 32;
- clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_local_blocks_fetched = 25;
- clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_blocks_fetched = 29;
- clearShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_blocks_fetched = 31;
- clearShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_bytes_read = 33;
- clearShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_bytes_read = 35;
- clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_fetch_fallback_count = 34;
- clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_fetch_fallback_count = 54;
- clearShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_fetch_fallback_count = 43;
- clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_blocks_fetched = 36;
- clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_blocks_fetched = 56;
- clearShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_blocks_fetched = 45;
- clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_bytes_read = 40;
- clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_bytes_read = 60;
- clearShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_bytes_read = 49;
- clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_chunks_fetched = 38;
- clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_chunks_fetched = 58;
- clearShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_chunks_fetched = 47;
- clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_blocks_fetched = 35;
- clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 55;
- clearShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 44;
- clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_bytes_read = 39;
- clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_bytes_read = 59;
- clearShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_bytes_read = 48;
- clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_chunks_fetched = 37;
- clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 57;
- clearShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 46;
- clearShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_req_duration = 51;
- clearShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_reqs_duration = 42;
- clearShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_reqs_duration = 62;
- clearShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 shuffle_mergers_count = 64;
- clearShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- clearShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- clearShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- clearShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read = 9;
- clearShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_read_bytes = 22;
- clearShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_bytes = 34;
- clearShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- clearShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read_records = 10;
- clearShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_records = 35;
- clearShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_records_read = 23;
- clearShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_read = 36;
- clearShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_written = 39;
- clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_blocks_fetched = 24;
- clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_blocks_fetched = 28;
- clearShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_blocks_fetched = 30;
- clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read = 27;
- clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read = 31;
- clearShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read = 33;
- clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read_to_disk = 28;
- clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 32;
- clearShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 34;
- clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_reqs_duration = 41;
- clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_reqs_duration = 61;
- clearShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_reqs_duration = 50;
- clearShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_total_blocks_fetched = 29;
- clearShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- clearShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write = 11;
- clearShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_bytes = 30;
- clearShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_bytes = 36;
- clearShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- clearShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_records = 31;
- clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write_records = 12;
- clearShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_records = 38;
- clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_time = 32;
- clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_time = 37;
- clearShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_write_time = 38;
- clearSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- clearSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- clearSources() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- clearSparkProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- clearSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string spark_user = 6;
- clearSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- clearSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
bool speculative = 12;
- clearSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool speculative = 12;
- clearSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
optional int64 sql_execution_id = 3;
- clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int32 stage_attempt_id = 2;
- clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- clearStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 stage_attempt_id = 41;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 stage_id = 1;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
int64 stage_id = 1;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 stage_id = 2;
- clearStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 stage_id = 40;
- clearStageIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- clearStageIds() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- clearStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- clearStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string start_offset = 2;
- clearStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 start_time = 2;
- clearStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
int64 start_timestamp = 6;
- clearStateOperators() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string status = 10;
- clearStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string status = 10;
- clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string storage_level = 2;
- clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string storage_level = 5;
- clearStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string storage_level = 4;
- clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 submission_time = 4;
- clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 submission_time = 8;
- clearSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 submission_time = 10;
- clearSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- clearSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 succeeded_tasks = 3;
- clearSystemProperties() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- clearTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 task_count = 4;
- clearTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 task_id = 1;
- clearTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 task_id = 1;
- clearTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string task_locality = 11;
- clearTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string task_locality = 11;
- clearTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- clearTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- clearTaskResourceRequests() - Method in class org.apache.spark.resource.ResourceProfileBuilder
-
- clearTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clearTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clearTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- clearTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 task_time = 1;
- clearThreshold() - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
-
Clears the threshold so that predict
will output raw prediction scores.
- clearThreshold() - Method in class org.apache.spark.mllib.classification.SVMModel
-
Clears the threshold so that predict
will output raw prediction scores.
- clearTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string timestamp = 4;
- clearToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 to_id = 2;
- clearToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 to_id = 2;
- clearTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- clearTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_cores = 7;
- clearTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int32 total_cores = 4;
- clearTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_duration = 13;
- clearTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_gc_time = 14;
- clearTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_input_bytes = 15;
- clearTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_off_heap_storage_memory = 4;
- clearTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_on_heap_storage_memory = 3;
- clearTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_read = 16;
- clearTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_write = 17;
- clearTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_tasks = 12;
- clearUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string update = 3;
- clearUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_disk = 6;
- clearUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_off_heap_storage_memory = 2;
- clearUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_on_heap_storage_memory = 1;
- clearUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_memory = 5;
- clearValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string value = 4;
- clearValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value1 = 1;
- clearValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value2 = 2;
- clearVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string vendor = 4;
- clearWrapper() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- clearWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- clearWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- clearWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- clearWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 write_time = 2;
- Clock - Interface in org.apache.spark.util
-
An interface to represent clocks, so that they can be mocked out in unit tests.
- CLogLog$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
-
- clone() - Method in class org.apache.spark.SparkConf
-
Copy this object
- clone() - Method in class org.apache.spark.sql.ExperimentalMethods
-
- clone() - Method in class org.apache.spark.sql.types.Decimal
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- clone() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- clone() - Method in class org.apache.spark.storage.StorageLevel
-
- clone() - Method in class org.apache.spark.util.random.BernoulliCellSampler
-
- clone() - Method in class org.apache.spark.util.random.BernoulliSampler
-
- clone() - Method in class org.apache.spark.util.random.PoissonSampler
-
- clone() - Method in interface org.apache.spark.util.random.RandomSampler
-
return a copy of the RandomSampler object
- clone(T, SerializerInstance, ClassTag<T>) - Static method in class org.apache.spark.util.Utils
-
Clone an object using a Spark serializer.
- cloneComplement() - Method in class org.apache.spark.util.random.BernoulliCellSampler
-
Return a sampler that is the complement of the range specified of the current sampler.
- cloneProperties(Properties) - Static method in class org.apache.spark.util.Utils
-
Create a new properties object with the same values as `props`
- close() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- close() - Method in class org.apache.spark.io.NioBufferedFileInputStream
-
- close() - Method in class org.apache.spark.io.ReadAheadInputStream
-
- close() - Method in interface org.apache.spark.security.CryptoStreamUtils.BaseErrorHandler
-
- close() - Method in class org.apache.spark.serializer.DeserializationStream
-
- close() - Method in class org.apache.spark.serializer.SerializationStream
-
- close(Throwable) - Method in class org.apache.spark.sql.ForeachWriter
-
Called when stopping to process one partition of new data in the executor side.
- close() - Method in class org.apache.spark.sql.SparkSession
-
Synonym for stop()
.
- close() - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- close() - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
-
Called to close all the columns in this batch.
- close() - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Cleans up memory for this column vector.
- close() - Method in class org.apache.spark.storage.BufferReleasingInputStream
-
- close() - Method in class org.apache.spark.storage.CountingWritableChannel
-
- close() - Method in class org.apache.spark.storage.TimeTrackingOutputStream
-
- close() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
- close() - Method in class org.apache.spark.streaming.util.WriteAheadLog
-
Close this log and release any resources.
- ClosureCleaner - Class in org.apache.spark.util
-
A cleaner that renders closures serializable if they can be done so safely.
- ClosureCleaner() - Constructor for class org.apache.spark.util.ClosureCleaner
-
- closureSerializer() - Method in class org.apache.spark.SparkEnv
-
- cls() - Method in class org.apache.spark.sql.types.ObjectType
-
- cls() - Method in class org.apache.spark.util.MethodIdentifier
-
- clsTag() - Method in interface org.apache.spark.sql.Encoder
-
A ClassTag that can be used to construct an Array to contain a collection of T
.
- cluster() - Method in class org.apache.spark.ml.clustering.ClusteringSummary
-
- cluster() - Method in class org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
-
- Cluster$() - Constructor for class org.apache.spark.mllib.clustering.KMeansModel.Cluster$
-
- CLUSTER_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- clusterCenter() - Method in class org.apache.spark.ml.clustering.ClusterData
-
- clusterCenters() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- clusterCenters() - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- clusterCenters() - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Leaf cluster centers.
- clusterCenters() - Method in class org.apache.spark.mllib.clustering.KMeansModel
-
- clusterCenters() - Method in class org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- ClusterData - Class in org.apache.spark.ml.clustering
-
Helper class for storing model data
- ClusterData(int, Vector) - Constructor for class org.apache.spark.ml.clustering.ClusterData
-
- clustered(Expression[]) - Static method in class org.apache.spark.sql.connector.distributions.Distributions
-
Creates a distribution where tuples that share the same values for clustering expressions are
co-located in the same partition.
- clustered(Expression[]) - Static method in class org.apache.spark.sql.connector.distributions.LogicalDistributions
-
- ClusteredDistribution - Interface in org.apache.spark.sql.connector.distributions
-
A distribution where tuples that share the same values for clustering expressions are co-located
in the same partition.
- clusterIdx() - Method in class org.apache.spark.ml.clustering.ClusterData
-
- clustering() - Method in interface org.apache.spark.sql.connector.distributions.ClusteredDistribution
-
Returns clustering expressions.
- ClusteringEvaluator - Class in org.apache.spark.ml.evaluation
-
Evaluator for clustering results.
- ClusteringEvaluator(String) - Constructor for class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- ClusteringEvaluator() - Constructor for class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- ClusteringMetrics - Class in org.apache.spark.ml.evaluation
-
Metrics for clustering, which expects two input columns: prediction and label.
- ClusteringSummary - Class in org.apache.spark.ml.clustering
-
Summary of clustering algorithms.
- CLUSTERS_CONFIG_PREFIX() - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
-
- clusterSchedulerError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- clusterSizes() - Method in class org.apache.spark.ml.clustering.ClusteringSummary
-
- ClusterStats(Vector, double, double) - Constructor for class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
-
- ClusterStats$() - Constructor for class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats$
-
- clusterWeights() - Method in class org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- cmdOnlyWorksOnPartitionedTablesError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cmdOnlyWorksOnTableWithLocationError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- cn() - Method in class org.apache.spark.mllib.feature.VocabWord
-
- coalesce(int) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int) - Method in class org.apache.spark.api.java.JavaRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int, boolean) - Method in class org.apache.spark.api.java.JavaRDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int, RDD<?>) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
Runs the packing algorithm and returns an array of PartitionGroups that if possible are
load balanced and grouped by locality
- coalesce(int, RDD<?>) - Method in interface org.apache.spark.rdd.PartitionCoalescer
-
Coalesce the partitions of the given RDD.
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
-
Return a new RDD that is reduced into numPartitions
partitions.
- coalesce(int) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset that has exactly numPartitions
partitions, when the fewer partitions
are requested.
- coalesce(Column...) - Static method in class org.apache.spark.sql.functions
-
Returns the first column that is not null, or null if all inputs are null.
- coalesce(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Returns the first column that is not null, or null if all inputs are null.
- CoarseGrainedClusterMessage - Interface in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages
-
- CoarseGrainedClusterMessages.AddWebUIFilter - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.AddWebUIFilter$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.DecommissionExecutor$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.DecommissionExecutorsOnHost - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.DecommissionExecutorsOnHost$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ExecutorDecommissioning - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ExecutorDecommissioning$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ExecutorDecommissionSigReceived$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.GetExecutorLossReason - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.GetExecutorLossReason$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.IsExecutorAlive - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.IsExecutorAlive$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillExecutors - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillExecutors$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillExecutorsOnHost - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillExecutorsOnHost$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillTask - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.KillTask$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.LaunchedExecutor - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.LaunchedExecutor$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.LaunchTask - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.LaunchTask$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.MiscellaneousProcessAdded - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.MiscellaneousProcessAdded$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RegisterClusterManager - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RegisterClusterManager$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RegisterExecutor - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RegisterExecutor$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RemoveExecutor - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RemoveExecutor$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RemoveWorker - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RemoveWorker$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RequestExecutors - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RequestExecutors$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RetrieveDelegationTokens$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RetrieveLastAllocatedExecutorId$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RetrieveSparkAppConfig - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.RetrieveSparkAppConfig$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ReviveOffers$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.SetupDriver - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.SetupDriver$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ShufflePushCompletion - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.ShufflePushCompletion$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.Shutdown - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.Shutdown$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.SparkAppConfig - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.SparkAppConfig$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.StatusUpdate - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.StatusUpdate$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.StopDriver$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.StopExecutor$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.StopExecutors$ - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.UpdateDelegationTokens - Class in org.apache.spark.scheduler.cluster
-
- CoarseGrainedClusterMessages.UpdateDelegationTokens$ - Class in org.apache.spark.scheduler.cluster
-
- code() - Method in class org.apache.spark.mllib.feature.VocabWord
-
- CodegenMetrics - Class in org.apache.spark.metrics.source
-
Metrics for code generation.
- CodegenMetrics() - Constructor for class org.apache.spark.metrics.source.CodegenMetrics
-
- codeLen() - Method in class org.apache.spark.mllib.feature.VocabWord
-
- coefficientMatrix() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- coefficients() - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
- coefficients() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
A vector of model coefficients for "binomial" logistic regression.
- coefficients() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- coefficients() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- coefficients() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- coefficientStandardErrors() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
- coefficientStandardErrors() - Method in class org.apache.spark.ml.regression.LinearRegressionSummary
-
- cogroup(JavaPairRDD<K, W>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(JavaPairRDD<K, W>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(RDD<Tuple2<K, W>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(RDD<Tuple2<K, W>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(RDD<Tuple2<K, W>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other
, return a resulting RDD that contains a tuple with the
list of values for that key in this
as well as other
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
, return a resulting RDD that contains a
tuple with the list of values for that key in this
, other1
and other2
.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this
or other1
or other2
or other3
,
return a resulting RDD that contains a tuple with the list of values
for that key in this
, other1
, other2
and other3
.
- cogroup(KeyValueGroupedDataset<K, U>, Function3<K, Iterator<V>, Iterator<U>, TraversableOnce<R>>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Applies the given function to each cogrouped data.
- cogroup(KeyValueGroupedDataset<K, U>, CoGroupFunction<K, V, U, R>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Applies the given function to each cogrouped data.
- cogroup(JavaPairDStream<K, W>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- cogroup(JavaPairDStream<K, W>, int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- cogroup(JavaPairDStream<K, W>, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- cogroup(DStream<Tuple2<K, W>>, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- cogroup(DStream<Tuple2<K, W>>, int, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- cogroup(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this
DStream and other
DStream.
- CoGroupedRDD<K> - Class in org.apache.spark.rdd
-
:: DeveloperApi ::
An RDD that cogroups its parents.
- CoGroupedRDD(Seq<RDD<? extends Product2<K, ?>>>, Partitioner, ClassTag<K>) - Constructor for class org.apache.spark.rdd.CoGroupedRDD
-
- CoGroupFunction<K,V1,V2,R> - Interface in org.apache.spark.api.java.function
-
A function that returns zero or more output records from each grouping key and its values from 2
Datasets.
- cogroupSorted(KeyValueGroupedDataset<K, U>, Seq<Column>, Seq<Column>, Function3<K, Iterator<V>, Iterator<U>, TraversableOnce<R>>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Applies the given function to each sorted cogrouped data.
- cogroupSorted(KeyValueGroupedDataset<K, U>, Column[], Column[], CoGroupFunction<K, V, U, R>, Encoder<R>) - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Applies the given function to each sorted cogrouped data.
- col(String) - Method in class org.apache.spark.sql.Dataset
-
Selects column based on the column name and returns it as a
Column
.
- col(String) - Static method in class org.apache.spark.sql.functions
-
Returns a
Column
based on the given column name.
- COL_POS_KEY() - Static method in class org.apache.spark.sql.Dataset
-
- coldStartStrategy() - Method in class org.apache.spark.ml.recommendation.ALS
-
- coldStartStrategy() - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- coldStartStrategy() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
-
Param for strategy for dealing with unknown or new users/items at prediction time.
- colIter() - Method in class org.apache.spark.ml.linalg.DenseMatrix
-
- colIter() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Returns an iterator of column vectors.
- colIter() - Method in class org.apache.spark.ml.linalg.SparseMatrix
-
- colIter() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
-
- colIter() - Method in interface org.apache.spark.mllib.linalg.Matrix
-
Returns an iterator of column vectors.
- colIter() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
-
- collect() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return an array that contains all of the elements in this RDD.
- collect() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
- collect() - Method in class org.apache.spark.rdd.RDD
-
Return an array that contains all of the elements in this RDD.
- collect(PartialFunction<T, U>, ClassTag<U>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD that contains all matching values by applying f
.
- collect() - Method in class org.apache.spark.sql.Dataset
-
Returns an array that contains all rows in this Dataset.
- collect_list(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns a list of objects with duplicates.
- collect_list(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns a list of objects with duplicates.
- collect_set(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns a set of objects with duplicate elements eliminated.
- collect_set(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns a set of objects with duplicate elements eliminated.
- collectAsList() - Method in class org.apache.spark.sql.Dataset
-
Returns a Java list that contains all rows in this Dataset.
- collectAsMap() - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return the key-value pairs in this RDD to the master as a Map.
- collectAsMap() - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return the key-value pairs in this RDD to the master as a Map.
- collectAsync() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of collect
, which returns a future for
retrieving an array containing all of the elements in this RDD.
- collectAsync() - Method in class org.apache.spark.rdd.AsyncRDDActions
-
Returns a future for retrieving all elements of this RDD.
- collectEdges(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
-
Returns an RDD that contains for each vertex v its local edges,
i.e., the edges that are incident on v, in the user-specified direction.
- collectionAccumulator() - Method in class org.apache.spark.SparkContext
-
Create and register a CollectionAccumulator
, which starts with empty list and accumulates
inputs by adding them into the list.
- collectionAccumulator(String) - Method in class org.apache.spark.SparkContext
-
Create and register a CollectionAccumulator
, which starts with empty list and accumulates
inputs by adding them into the list.
- CollectionAccumulator<T> - Class in org.apache.spark.util
-
- CollectionAccumulator() - Constructor for class org.apache.spark.util.CollectionAccumulator
-
- CollectionsUtils - Class in org.apache.spark.util
-
- CollectionsUtils() - Constructor for class org.apache.spark.util.CollectionsUtils
-
- collectNeighborIds(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
-
Collect the neighbor vertex ids for each vertex.
- collectNeighbors(EdgeDirection) - Method in class org.apache.spark.graphx.GraphOps
-
Collect the neighbor vertex attributes for each vertex.
- collectPartitions(int[]) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return an array that contains all of the elements in a specific partition of this RDD.
- collectSubModels() - Method in interface org.apache.spark.ml.param.shared.HasCollectSubModels
-
Param for whether to collect a list of sub-models trained during tuning.
- collectSubModels() - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- collectSubModels() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- colPtrs() - Method in class org.apache.spark.ml.linalg.SparseMatrix
-
- colPtrs() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
-
- colRegex(String) - Method in class org.apache.spark.sql.Dataset
-
Selects column based on the column name specified as a regex and returns it as
Column
.
- colsPerBlock() - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
- colStats(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Computes column-wise summary statistics for the input RDD[Vector].
- Column - Class in org.apache.spark.sql.catalog
-
A column in Spark, as returned by
listColumns
method in
Catalog
.
- Column(String, String, String, boolean, boolean, boolean) - Constructor for class org.apache.spark.sql.catalog.Column
-
- Column - Class in org.apache.spark.sql
-
A column that will be computed based on the data in a DataFrame
.
- Column(Expression) - Constructor for class org.apache.spark.sql.Column
-
- Column(String) - Constructor for class org.apache.spark.sql.Column
-
- Column - Interface in org.apache.spark.sql.connector.catalog
-
An interface representing a column of a
Table
.
- column() - Method in class org.apache.spark.sql.connector.catalog.TableChange.After
-
- column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Avg
-
- column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Count
-
- column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Max
-
- column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Min
-
- column() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Sum
-
- column(String) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
-
Create a named reference expression for a (nested) column.
- column(String) - Static method in class org.apache.spark.sql.functions
-
Returns a
Column
based on the given column name.
- column(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
-
Returns the column at `ordinal`.
- columnAliases() - Method in interface org.apache.spark.sql.connector.catalog.View
-
The view column aliases.
- columnAliasInOperationNotAllowedError(String, SqlBaseParser.TableAliasContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- columnAlreadyExistsError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ColumnarArray - Class in org.apache.spark.sql.vectorized
-
- ColumnarArray(ColumnVector, int, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarArray
-
- ColumnarBatch - Class in org.apache.spark.sql.vectorized
-
This class wraps multiple ColumnVectors as a row-wise table.
- ColumnarBatch(ColumnVector[]) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatch
-
- ColumnarBatch(ColumnVector[], int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatch
-
Create a new batch from existing column vectors.
- ColumnarBatchRow - Class in org.apache.spark.sql.vectorized
-
This class wraps an array of
ColumnVector
and provides a row view.
- ColumnarBatchRow(ColumnVector[]) - Constructor for class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- ColumnarMap - Class in org.apache.spark.sql.vectorized
-
- ColumnarMap(ColumnVector, ColumnVector, int, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarMap
-
- ColumnarRow - Class in org.apache.spark.sql.vectorized
-
- ColumnarRow(ColumnVector, int) - Constructor for class org.apache.spark.sql.vectorized.ColumnarRow
-
- columnarSupportMode() - Method in interface org.apache.spark.sql.connector.read.Scan
-
Subclasses can implement this method to indicate if the support for columnar data should
be determined by each partition or is set as a default for the whole scan.
- columnComments() - Method in interface org.apache.spark.sql.connector.catalog.View
-
The view column comments.
- ColumnDefaultValue - Class in org.apache.spark.sql.connector.catalog
-
A class representing the default value of a column.
- ColumnDefaultValue(String, Literal<?>) - Constructor for class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
-
- ColumnName - Class in org.apache.spark.sql
-
A convenient class used for constructing schema.
- ColumnName(String) - Constructor for class org.apache.spark.sql.ColumnName
-
- columnNotDefinedInTableError(String, String, String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnNotFoundError(String) - Method in interface org.apache.spark.sql.errors.CompilationErrors
-
- columnNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnNotFoundInExistingColumnsError(String, String, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnNotFoundInSchemaError(StructField, Option<StructType>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnNotInGroupByClauseError(Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnProperties() - Method in class org.apache.spark.sql.connector.catalog.index.TableIndex
-
- ColumnPruner - Class in org.apache.spark.ml.feature
-
Utility transformer for removing temporary columns from a DataFrame.
- ColumnPruner(String, Set<String>) - Constructor for class org.apache.spark.ml.feature.ColumnPruner
-
- ColumnPruner(Set<String>) - Constructor for class org.apache.spark.ml.feature.ColumnPruner
-
- columns() - Method in class org.apache.spark.sql.connector.catalog.index.TableIndex
-
- columns() - Method in interface org.apache.spark.sql.connector.catalog.Table
-
Returns the columns of this table.
- columns() - Method in class org.apache.spark.sql.Dataset
-
Returns all column names as an array.
- columnSchema() - Static method in class org.apache.spark.ml.image.ImageSchema
-
Schema for the image column: Row(String, Int, Int, Int, Int, Array[Byte])
- ColumnsHelper(Column[]) - Constructor for class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.ColumnsHelper
-
- columnSimilarities() - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Compute all cosine similarities between columns of this matrix using the brute-force
approach of computing normalized dot products.
- columnSimilarities() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Compute all cosine similarities between columns of this matrix using the brute-force
approach of computing normalized dot products.
- columnSimilarities(double) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Compute similarities between columns of this matrix using a sampling approach.
- ColumnStatistics - Interface in org.apache.spark.sql.connector.read.colstats
-
An interface to represent column statistics, which is part of
Statistics
.
- columnStatisticsDeserializationNotSupportedError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnStatisticsSerializationNotSupportedError(String, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- columnStats() - Method in interface org.apache.spark.sql.connector.read.Statistics
-
- columnsToPrune() - Method in class org.apache.spark.ml.feature.ColumnPruner
-
- columnToOldVector(Dataset<?>, String) - Static method in class org.apache.spark.ml.util.DatasetUtils
-
- columnToVector(Dataset<?>, String) - Static method in class org.apache.spark.ml.util.DatasetUtils
-
Cast a column in a Dataset to Vector type.
- columnTypeNotSupportStatisticsCollectionError(String, TableIdentifier, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- ColumnVector - Class in org.apache.spark.sql.vectorized
-
An interface representing in-memory columnar data in Spark.
- combinationQueryResultClausesUnsupportedError(SqlBaseParser.QueryOrganizationContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Simplified version of combineByKey that hash-partitions the output RDD and uses map-side
aggregation.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Simplified version of combineByKey that hash-partitions the resulting RDD using the existing
partitioner/parallelism level and using map-side aggregation.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the
existing partitioner/parallelism level.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Combine elements of each key in DStream's RDDs using custom function.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Combine elements of each key in DStream's RDDs using custom function.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, ClassTag<C>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Combine elements of each key in DStream's RDDs using custom functions.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, ClassTag<C>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the
existing partitioner/parallelism level.
- combineCombinersByKey(Iterator<? extends Product2<K, C>>, TaskContext) - Method in class org.apache.spark.Aggregator
-
- combineValuesByKey(Iterator<? extends Product2<K, V>>, TaskContext) - Method in class org.apache.spark.Aggregator
-
- command() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperation
-
Returns the SQL command that is being performed.
- command() - Method in interface org.apache.spark.sql.connector.write.RowLevelOperationInfo
-
Returns the row-level SQL command (e.g.
- commandExecutionInRunnerUnsupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- CommandLineLoggingUtils - Interface in org.apache.spark.util
-
- CommandLineUtils - Interface in org.apache.spark.util
-
Contains basic command line parsing functionality and methods to parse some common Spark CLI
options.
- commandNotSupportNestedColumnError(String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- commandUnsupportedInV2TableError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- comment() - Method in interface org.apache.spark.sql.connector.catalog.Column
-
Returns the comment of this table column.
- comment() - Method in interface org.apache.spark.sql.connector.catalog.MetadataColumn
-
Documentation for this metadata column, or null.
- comment() - Method in class org.apache.spark.sql.connector.catalog.TableChange.AddColumn
-
- commentOnTableUnsupportedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- commit(Function0<Parsers.Parser<T>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- commit(Offset) - Method in interface org.apache.spark.sql.connector.read.streaming.SparkDataStream
-
Informs the source that Spark has completed processing all data for offsets less than or
equal to `end` and will only request offsets greater than `end` in the future.
- commit(WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
-
Commits this writing job with a list of commit messages.
- commit() - Method in interface org.apache.spark.sql.connector.write.DataWriter
-
- commit(long, WriterCommitMessage[]) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
-
Commits this writing job for the specified epoch with a list of commit messages.
- COMMIT_TIME_MS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- commitAllPartitions(long[]) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
-
- commitDeniedError(int, long, int, int, int) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- commitStagedChanges() - Method in interface org.apache.spark.sql.connector.catalog.StagedTable
-
Finalize the creation or replacement of this table.
- commitTask(OutputCommitter, TaskAttemptContext, int, int) - Static method in class org.apache.spark.mapred.SparkHadoopMapRedUtil
-
Commits a task output.
- commitTimeMs() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
-
- commonHeaderNodes(HttpServletRequest) - Static method in class org.apache.spark.ui.UIUtils
-
- comparator(Schedulable, Schedulable) - Method in interface org.apache.spark.scheduler.SchedulingAlgorithm
-
- comparatorReturnsNull(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- compare(PartitionGroup, PartitionGroup) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer.partitionGroupOrdering$
-
- compare(byte, byte) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
-
- compare(Decimal) - Method in class org.apache.spark.sql.types.Decimal
-
- compare(Decimal, Decimal) - Method in interface org.apache.spark.sql.types.Decimal.DecimalIsConflicted
-
- compare(Decimal, Decimal) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
-
- compare(double, double) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
-
- compare(double, double) - Method in class org.apache.spark.sql.types.DoubleType.DoubleAsIfIntegral$
-
- compare(float, float) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
-
- compare(float, float) - Method in class org.apache.spark.sql.types.FloatType.FloatAsIfIntegral$
-
- compare(int, int) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
-
- compare(long, long) - Static method in class org.apache.spark.sql.types.LongExactNumeric
-
- compare(short, short) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
-
- compare(RDDInfo) - Method in class org.apache.spark.storage.RDDInfo
-
- compareTo(NumericHistogram.Coord) - Method in class org.apache.spark.sql.util.NumericHistogram.Coord
-
- compareTo(SparkShutdownHook) - Method in class org.apache.spark.util.SparkShutdownHook
-
- CompilationErrors - Interface in org.apache.spark.sql.errors
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- compileAggregate(AggregateFunc) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- compileAggregate(AggregateFunc) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- compileExpression(Expression) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Converts V2 expression to String representing a SQL expression.
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- compileExpression(Expression) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- compilerError(CompileException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- compileValue(Object) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Converts value to SQL expression.
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- compileValue(Object) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- Complete() - Static method in class org.apache.spark.sql.streaming.OutputMode
-
OutputMode in which all the rows in the streaming DataFrame/Dataset will be written
to the sink every time there are some updates.
- completed() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- COMPLETED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- COMPLETED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- completedIndices() - Method in class org.apache.spark.status.LiveJob
-
- completedIndices() - Method in class org.apache.spark.status.LiveStage
-
- completedStages() - Method in class org.apache.spark.status.LiveJob
-
- completedTasks() - Method in class org.apache.spark.status.api.v1.ExecutorSummary
-
- completedTasks() - Method in class org.apache.spark.status.LiveJob
-
- completedTasks() - Method in class org.apache.spark.status.LiveStage
-
- COMPLETION_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- COMPLETION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- completionTime() - Method in class org.apache.spark.scheduler.StageInfo
-
Time when the stage completed or when the stage was cancelled.
- completionTime() - Method in class org.apache.spark.status.api.v1.JobData
-
- completionTime() - Method in class org.apache.spark.status.api.v1.StageData
-
- completionTime() - Method in class org.apache.spark.status.LiveJob
-
- ComplexFutureAction<T> - Class in org.apache.spark
-
A
FutureAction
for actions that could trigger multiple Spark jobs.
- ComplexFutureAction(Function1<JobSubmitter, Future<T>>) - Constructor for class org.apache.spark.ComplexFutureAction
-
- componentName() - Method in class org.apache.spark.resource.ResourceID
-
- compositeLimit(ReadLimit[]) - Static method in interface org.apache.spark.sql.connector.read.streaming.ReadLimit
-
- CompositeReadLimit - Class in org.apache.spark.sql.connector.read.streaming
-
/**
Represents a
ReadLimit
where the
MicroBatchStream
should scan approximately
given maximum number of rows with at least the given minimum number of rows.
- compressed() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense column major, dense row major, sparse row major, or sparse column
major format, whichever uses less storage.
- compressed() - Method in interface org.apache.spark.ml.linalg.Vector
-
Returns a vector in either dense or sparse format, whichever uses less storage.
- compressed() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Returns a vector in either dense or sparse format, whichever uses less storage.
- compressedColMajor() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense or sparse column major format, whichever uses less storage.
- compressedContinuousInputStream(InputStream) - Method in interface org.apache.spark.io.CompressionCodec
-
- compressedContinuousInputStream(InputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
-
- compressedContinuousOutputStream(OutputStream) - Method in interface org.apache.spark.io.CompressionCodec
-
- compressedInputStream(InputStream) - Method in interface org.apache.spark.io.CompressionCodec
-
- compressedInputStream(InputStream) - Method in class org.apache.spark.io.LZ4CompressionCodec
-
- compressedInputStream(InputStream) - Method in class org.apache.spark.io.LZFCompressionCodec
-
- compressedInputStream(InputStream) - Method in class org.apache.spark.io.SnappyCompressionCodec
-
- compressedInputStream(InputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
-
- compressedOutputStream(OutputStream) - Method in interface org.apache.spark.io.CompressionCodec
-
- compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.LZ4CompressionCodec
-
- compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.LZFCompressionCodec
-
- compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.SnappyCompressionCodec
-
- compressedOutputStream(OutputStream) - Method in class org.apache.spark.io.ZStdCompressionCodec
-
- compressedRowMajor() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense or sparse row major format, whichever uses less storage.
- CompressionCodec - Interface in org.apache.spark.io
-
:: DeveloperApi ::
CompressionCodec allows the customization of choosing different compression implementations
to be used in block storage.
- compute(Partition, TaskContext) - Method in class org.apache.spark.api.r.BaseRRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.graphx.EdgeRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.graphx.VertexRDD
-
Provides the RDD[(VertexId, VD)]
equivalent output.
- compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.Gradient
-
Compute the gradient and loss given the features of a single data point.
- compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.Gradient
-
Compute the gradient and loss given the features of a single data point,
add the gradient to a provided vector to avoid creating new objects, and return loss.
- compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.HingeGradient
-
- compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.HingeGradient
-
- compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.L1Updater
-
- compute(Vector, double, Vector) - Method in class org.apache.spark.mllib.optimization.LeastSquaresGradient
-
- compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.LeastSquaresGradient
-
- compute(Vector, double, Vector, Vector) - Method in class org.apache.spark.mllib.optimization.LogisticGradient
-
- compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.SimpleUpdater
-
- compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.SquaredL2Updater
-
- compute(Vector, Vector, double, int, double) - Method in class org.apache.spark.mllib.optimization.Updater
-
Compute an updated value for weights given the gradient, stepSize, iteration number and
regularization parameter.
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.CoGroupedRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.HadoopRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.JdbcRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.NewHadoopRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.PartitionPruningRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.RDD
-
:: DeveloperApi ::
Implemented by subclasses to compute a given partition.
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.ShuffledRDD
-
- compute(Partition, TaskContext) - Method in class org.apache.spark.rdd.UnionRDD
-
- compute(Time) - Method in class org.apache.spark.streaming.api.java.JavaDStream
-
Generate an RDD for the given duration
- compute(Time) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Method that generates an RDD for the given Duration
- compute(Time) - Method in class org.apache.spark.streaming.dstream.ConstantInputDStream
-
- compute(Time) - Method in class org.apache.spark.streaming.dstream.DStream
-
Method that generates an RDD for the given time
- compute(Time) - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- compute(long, long, long, long) - Method in interface org.apache.spark.streaming.scheduler.rate.RateEstimator
-
Computes the number of records the stream attached to this RateEstimator
should ingest per second, given an update on the size and completion
times of the latest batch.
- computeClusterStats(Dataset<Row>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
-
The method takes the input dataset and computes the aggregated values
about a cluster which are needed by the algorithm.
- computeClusterStats(Dataset<Row>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
The method takes the input dataset and computes the aggregated values
about a cluster which are needed by the algorithm.
- computeColumnSummaryStatistics() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes column-wise summary statistics.
- computeCorrelation(RDD<Object>, RDD<Object>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
-
Compute correlation for two datasets.
- computeCorrelation(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation for two datasets.
- computeCorrelation(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
Compute Spearman's correlation for two datasets.
- computeCorrelationMatrix(RDD<Vector>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
-
Compute the correlation matrix S, for the input matrix, where S(i, j) is the correlation
between column i and j.
- computeCorrelationMatrix(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation matrix S, for the input matrix, where S(i, j) is the
correlation between column i and j.
- computeCorrelationMatrix(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
Compute Spearman's correlation matrix S, for the input matrix, where S(i, j) is the
correlation between column i and j.
- computeCorrelationMatrixFromCovariance(Matrix) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation matrix from the covariance matrix.
- computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Method in interface org.apache.spark.mllib.stat.correlation.Correlation
-
Combine the two input RDD[Double]s into an RDD[Vector] and compute the correlation using the
correlation implementation for RDD[Vector].
- computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
- computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
- computeCost(Dataset<?>) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- computeCost(Vector) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Computes the squared distance between the input point and the cluster center it belongs to.
- computeCost(RDD<Vector>) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Computes the sum of squared distances between the input points and their corresponding cluster
centers.
- computeCost(JavaRDD<Vector>) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Java-friendly version of computeCost()
.
- computeCost(RDD<Vector>) - Method in class org.apache.spark.mllib.clustering.KMeansModel
-
Return the K-means cost (sum of squared distances of points to their nearest center) for this
model on the given data.
- computeCovariance() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the covariance matrix, treating each row as an observation.
- computeError(org.apache.spark.mllib.tree.model.TreeEnsembleModel, RDD<LabeledPoint>) - Method in interface org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate error of the base learner for the gradient boosting calculation.
- computeError(double, double) - Method in interface org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate loss when the predictions are already known.
- computeFractionForSampleSize(int, long, boolean) - Static method in class org.apache.spark.util.random.SamplingUtils
-
Returns a sampling rate that guarantees a sample of size greater than or equal to
sampleSizeLowerBound 99.99% of the time.
- computeGradient(DenseMatrix<Object>, DenseMatrix<Object>, Vector, int) - Method in interface org.apache.spark.ml.ann.TopologyModel
-
Computes gradient for the network
- computeGramianMatrix() - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Computes the Gramian matrix A^T A
.
- computeGramianMatrix() - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the Gramian matrix A^T A
.
- computeInitialPredictionAndError(RDD<TreePoint>, double, DecisionTreeRegressionModel, Loss, Broadcast<Split[][]>) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Compute the initial predictions and errors for a dataset for the first
iteration of gradient boosting.
- computeInitialPredictionAndError(RDD<LabeledPoint>, double, DecisionTreeModel, Loss) - Static method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
Compute the initial predictions and errors for a dataset for the first
iteration of gradient boosting.
- computePreferredLocations(Seq<InputFormatInfo>) - Static method in class org.apache.spark.scheduler.InputFormatInfo
-
Computes the preferred locations based on input(s) and returned a location to block map.
- computePrevDelta(DenseMatrix<Object>, DenseMatrix<Object>, DenseMatrix<Object>) - Method in interface org.apache.spark.ml.ann.LayerModel
-
Computes the delta for back propagation.
- computePrincipalComponents(int) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the top k principal components only.
- computePrincipalComponentsAndExplainedVariance(int) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the top k principal components and a vector of proportions of
variance explained by each principal component.
- computeProbability(double) - Method in interface org.apache.spark.mllib.tree.loss.ClassificationLoss
-
Computes the class probability given the margin.
- computeSilhouetteCoefficient(Broadcast<Map<Object, Tuple2<Vector, Object>>>, Vector, double, double) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
-
It computes the Silhouette coefficient for a point.
- computeSilhouetteCoefficient(Broadcast<Map<Object, SquaredEuclideanSilhouette.ClusterStats>>, Vector, double, double, double) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
It computes the Silhouette coefficient for a point.
- computeSilhouetteScore(Dataset<?>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.CosineSilhouette
-
Compute the Silhouette score of the dataset using the cosine distance measure.
- computeSilhouetteScore(Dataset<?>, String, String, String) - Static method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
Compute the Silhouette score of the dataset using squared Euclidean distance measure.
- computeStatisticsNotExpectedError(SqlBaseParser.IdentifierContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- computeSVD(int, boolean, double) - Method in class org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Computes the singular value decomposition of this IndexedRowMatrix.
- computeSVD(int, boolean, double) - Method in class org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes singular value decomposition of this matrix.
- computeThresholdByKey(Map<K, AcceptanceResult>, Map<K, Object>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Given the result returned by getCounts, determine the threshold for accepting items to
generate exact sample size.
- computeWeightedError(RDD<org.apache.spark.ml.feature.Instance>, DecisionTreeRegressionModel[], double[], Loss) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Method to calculate error of the base learner for the gradient boosting calculation.
- computeWeightedError(RDD<TreePoint>, RDD<Tuple2<Object, Object>>) - Static method in class org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Method to calculate error of the base learner for the gradient boosting calculation.
- concat(Column...) - Static method in class org.apache.spark.sql.functions
-
Concatenates multiple input columns together into a single column.
- concat(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Concatenates multiple input columns together into a single column.
- concat_ws(String, Column...) - Static method in class org.apache.spark.sql.functions
-
Concatenates multiple input string columns together into a single string column,
using the given separator.
- concat_ws(String, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Concatenates multiple input string columns together into a single string column,
using the given separator.
- concatArraysWithElementsExceedLimitError(long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- concurrentQueryInstanceError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- concurrentStreamLogUpdate(long) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- conf() - Method in interface org.apache.spark.api.plugin.PluginContext
-
Configuration of the Spark application.
- Conf(int, int, double, double, double, double, double, double) - Constructor for class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- conf() - Method in class org.apache.spark.SparkEnv
-
- conf() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- conf() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- conf() - Method in class org.apache.spark.sql.SparkSession
-
- confidence() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
-
Returns the confidence of the rule.
- confidence() - Method in class org.apache.spark.partial.BoundedDouble
-
- confidence() - Method in class org.apache.spark.util.sketch.CountMinSketch
-
- config(String, String) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, long) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, double) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, boolean) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(SparkConf) - Method in class org.apache.spark.sql.SparkSession.Builder
-
Sets a list of config options based on the given SparkConf
.
- configRemovedInVersionError(String, String, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- configTestLog4j2(String) - Static method in class org.apache.spark.TestUtils
-
config a log4j2 properties used for testsuite
- Configurable - Interface in org.apache.spark.input
-
A trait to implement Configurable
interface.
- configuration() - Method in class org.apache.spark.scheduler.InputFormatInfo
-
- CONFIGURATION_INSTANTIATION_LOCK() - Static method in class org.apache.spark.rdd.HadoopRDD
-
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
- CONFIGURATION_INSTANTIATION_LOCK() - Static method in class org.apache.spark.rdd.NewHadoopRDD
-
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
- conflictingAttributesInJoinConditionError(AttributeSet, LogicalPlan, LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- confusionMatrix() - Method in class org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns confusion matrix:
predicted classes are in columns,
they are ordered by class label ascending,
as in "labels"
- connectedComponents() - Method in class org.apache.spark.graphx.GraphOps
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- connectedComponents(int) - Method in class org.apache.spark.graphx.GraphOps
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- ConnectedComponents - Class in org.apache.spark.graphx.lib
-
Connected components algorithm.
- ConnectedComponents() - Constructor for class org.apache.spark.graphx.lib.ConnectedComponents
-
- consequent() - Method in class org.apache.spark.mllib.fpm.AssociationRules.Rule
-
- ConstantInputDStream<T> - Class in org.apache.spark.streaming.dstream
-
An input stream that always returns the same RDD on each time step.
- ConstantInputDStream(StreamingContext, RDD<T>, ClassTag<T>) - Constructor for class org.apache.spark.streaming.dstream.ConstantInputDStream
-
- constructorNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- constructTree(org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData[]) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
Given a list of nodes from a tree, construct the tree.
- constructTrees(RDD<org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData>) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- contains(Param<?>) - Method in class org.apache.spark.ml.param.ParamMap
-
Checks whether a parameter is explicitly specified.
- contains(String) - Method in class org.apache.spark.SparkConf
-
Does the configuration contain a given parameter?
- contains(Object) - Method in class org.apache.spark.sql.Column
-
Contains the other element.
- contains(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a boolean.
- contains(String) - Method in class org.apache.spark.sql.types.Metadata
-
Tests whether this Metadata contains a binding for a key.
- contains(T) - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- containsAttributes(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> attributes = 27;
- containsAttributes(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> attributes = 27;
- containsAttributes(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> attributes = 27;
- containsCustomMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
map<string, int64> custom_metrics = 12;
- containsCustomMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
map<string, int64> custom_metrics = 12;
- containsCustomMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
map<string, int64> custom_metrics = 12;
- containsDurationMs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, int64> duration_ms = 7;
- containsDurationMs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, int64> duration_ms = 7;
- containsDurationMs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, int64> duration_ms = 7;
- containsEventTime(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> event_time = 8;
- containsEventTime(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> event_time = 8;
- containsEventTime(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> event_time = 8;
- containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> executor_logs = 23;
- containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> executor_logs = 23;
- containsExecutorLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> executor_logs = 23;
- containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
map<string, string> executor_logs = 16;
- containsExecutorLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
map<string, string> executor_logs = 16;
- containsExecutorLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
map<string, string> executor_logs = 16;
- containsExecutorResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- containsExecutorResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- containsExecutorResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- containsExecutorSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- containsExecutorSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- containsExecutorSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- containsJobs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- containsJobs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- containsJobs(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- containsKey(Object) - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- containsKey(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- containsKilledTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, int32> killed_tasks_summary = 48;
- containsKilledTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, int32> killed_tasks_summary = 48;
- containsKilledTasksSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, int32> killed_tasks_summary = 48;
- containsKillTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
map<string, int32> kill_tasks_summary = 20;
- containsKillTasksSummary(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
map<string, int32> kill_tasks_summary = 20;
- containsKillTasksSummary(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
map<string, int32> kill_tasks_summary = 20;
- containsLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
map<string, int64> locality = 3;
- containsLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
map<string, int64> locality = 3;
- containsLocality(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
map<string, int64> locality = 3;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
map<string, int64> metrics = 1;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
map<string, int64> metrics = 1;
- containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
map<string, int64> metrics = 1;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
map<string, string> metrics = 3;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
map<string, string> metrics = 3;
- containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
map<string, string> metrics = 3;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
map<string, string> metrics = 8;
- containsMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
map<string, string> metrics = 8;
- containsMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
map<string, string> metrics = 8;
- containsMetricValues(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, string> metric_values = 14;
- containsMetricValues(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, string> metric_values = 14;
- containsMetricValues(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, string> metric_values = 14;
- containsModifiedConfigs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<string, string> modified_configs = 6;
- containsModifiedConfigs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<string, string> modified_configs = 6;
- containsModifiedConfigs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<string, string> modified_configs = 6;
- containsNaN() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- containsNull() - Method in class org.apache.spark.sql.types.ArrayType
-
- containsNull() - Method in class org.apache.spark.sql.util.SQLOpenHashSet
-
- containsObservedMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> observed_metrics = 12;
- containsObservedMetrics(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> observed_metrics = 12;
- containsObservedMetrics(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> observed_metrics = 12;
- containsProcessLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
map<string, string> process_logs = 7;
- containsProcessLogs(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
map<string, string> process_logs = 7;
- containsProcessLogs(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
map<string, string> process_logs = 7;
- containsResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- containsResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- containsResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- containsTaskResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- containsTaskResources(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- containsTaskResources(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- containsTasks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- containsTasks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- containsTasks(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- containsValue(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- contentType() - Method in class org.apache.spark.ui.JettyUtils.ServletParams
-
- context() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
- context() - Method in class org.apache.spark.ContextAwareIterator
-
- context() - Method in class org.apache.spark.InterruptibleIterator
-
- context() - Method in class org.apache.spark.rdd.RDD
-
- context() - Method in exception org.apache.spark.sql.AnalysisException
-
- context() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
- context() - Method in class org.apache.spark.streaming.dstream.DStream
-
Return the StreamingContext associated with this DStream
- ContextAwareIterator<T> - Class in org.apache.spark
-
:: DeveloperApi ::
A TaskContext aware iterator.
- ContextAwareIterator(TaskContext, Iterator<T>) - Constructor for class org.apache.spark.ContextAwareIterator
-
- ContextBarrierId - Class in org.apache.spark
-
For each barrier stage attempt, only at most one barrier() call can be active at any time, thus
we can use (stageId, stageAttemptId) to identify the stage attempt where the barrier() call is
from.
- ContextBarrierId(int, int) - Constructor for class org.apache.spark.ContextBarrierId
-
- Continuous() - Static method in class org.apache.spark.mllib.tree.configuration.FeatureType
-
- Continuous(long) - Static method in class org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
- Continuous(long, TimeUnit) - Static method in class org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
- Continuous(Duration) - Static method in class org.apache.spark.sql.streaming.Trigger
-
(Scala-friendly)
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
- Continuous(String) - Static method in class org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
- ContinuousPartitionReader<T> - Interface in org.apache.spark.sql.connector.read.streaming
-
A variation on
PartitionReader
for use with continuous streaming processing.
- ContinuousPartitionReaderFactory - Interface in org.apache.spark.sql.connector.read.streaming
-
- continuousProcessingUnsupportedByDataSourceError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- ContinuousSplit - Class in org.apache.spark.ml.tree
-
Split which tests a continuous feature.
- ContinuousStream - Interface in org.apache.spark.sql.connector.read.streaming
-
- conv(Column, int, int) - Static method in class org.apache.spark.sql.functions
-
Convert a number in a string column from one base to another.
- convert_timezone(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Converts the timestamp without time zone sourceTs
from the sourceTz
time zone to targetTz
.
- convert_timezone(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Converts the timestamp without time zone sourceTs
from the current time zone to targetTz
.
- convertCachedBatchToColumnarBatch(RDD<CachedBatch>, Seq<Attribute>, Seq<Attribute>, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Convert the cached data into a ColumnarBatch.
- convertCachedBatchToInternalRow(RDD<CachedBatch>, Seq<Attribute>, Seq<Attribute>, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Convert the cached batch into InternalRow
s.
- convertColumnarBatchToCachedBatch(RDD<ColumnarBatch>, Seq<Attribute>, StorageLevel, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Convert an RDD[ColumnarBatch]
into an RDD[CachedBatch]
in preparation for caching the data.
- convertHiveTableToCatalogTableError(SparkException, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- convertInternalRowToCachedBatch(RDD<InternalRow>, Seq<Attribute>, StorageLevel, SQLConf) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Convert an RDD[InternalRow]
into an RDD[CachedBatch]
in preparation for caching the data.
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Converts an instance of java.sql.Timestamp
to a custom java.sql.Timestamp
value.
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
java.sql timestamps are measured with millisecond accuracy (from Long.MinValue
milliseconds to Long.MaxValue milliseconds), while Spark timestamps are measured
at microseconds accuracy.
- convertJavaTimestampToTimestamp(Timestamp) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Convert java.sql.Timestamp to a LocalDateTime representing the same wall-clock time as the
value stored in a remote database.
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- convertJavaTimestampToTimestampNTZ(Timestamp) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- convertMatrixColumnsFromML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts matrix columns in an input Dataset to the
Matrix
type from the new
Matrix
type under the
spark.ml
package.
- convertMatrixColumnsFromML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts matrix columns in an input Dataset to the
Matrix
type from the new
Matrix
type under the
spark.ml
package.
- convertMatrixColumnsToML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts Matrix columns in an input Dataset from the
Matrix
type to the new
Matrix
type under the
spark.ml
package.
- convertMatrixColumnsToML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts Matrix columns in an input Dataset from the
Matrix
type to the new
Matrix
type under the
spark.ml
package.
- convertTableProperties(TableSpec) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Converts a LocalDateTime representing a TimestampNTZ type to an
instance of java.sql.Timestamp
.
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- convertTimestampNTZToJavaTimestamp(LocalDateTime) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- convertToCanonicalEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.GraphOps
-
Convert bi-directional edges into uni-directional ones.
- convertToOldLossType(String) - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
-
- convertToTimeUnit(long, TimeUnit) - Static method in class org.apache.spark.streaming.ui.UIUtils
-
Convert milliseconds
to the specified unit
.
- convertTransforms() - Method in class org.apache.spark.sql.connector.catalog.CatalogV2Implicits.TransformHelper
-
- convertVectorColumnsFromML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset to the
Vector
type from the new
Vector
type under the
spark.ml
package.
- convertVectorColumnsFromML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset to the
Vector
type from the new
Vector
type under the
spark.ml
package.
- convertVectorColumnsToML(Dataset<?>, String...) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset from the
Vector
type to the new
Vector
type under the
spark.ml
package.
- convertVectorColumnsToML(Dataset<?>, Seq<String>) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset from the
Vector
type to the new
Vector
type under the
spark.ml
package.
- Coord() - Constructor for class org.apache.spark.sql.util.NumericHistogram.Coord
-
- CoordinateMatrix - Class in org.apache.spark.mllib.linalg.distributed
-
Represents a matrix in coordinate format.
- CoordinateMatrix(RDD<MatrixEntry>, long, long) - Constructor for class org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
- CoordinateMatrix(RDD<MatrixEntry>) - Constructor for class org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Alternative constructor leaving matrix dimensions to be determined automatically.
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.FMClassificationModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.FMClassifier
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.LinearSVC
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.NaiveBayes
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.NaiveBayesModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.OneVsRestModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.DistributedLDAModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.KMeans
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.LDA
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.LocalLDAModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- copy(ParamMap) - Method in class org.apache.spark.ml.Estimator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.Evaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Binarizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.ChiSqSelectorModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.ColumnPruner
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.HashingTF
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.IDF
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.IDFModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Imputer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.ImputerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.IndexToString
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Interaction
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MaxAbsScaler
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MaxAbsScalerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinHashLSHModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinMaxScaler
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.PCA
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.PCAModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.PolynomialExpansion
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.RFormula
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.RFormulaModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.RobustScalerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.SQLTransformer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.StandardScaler
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.StandardScalerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Tokenizer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorAssembler
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorAttributeRewriter
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorIndexer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorIndexerModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- copy(ParamMap) - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- copy(ParamMap) - Method in class org.apache.spark.ml.fpm.FPGrowthModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- copy(Vector, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
-
y = x
- copy() - Method in class org.apache.spark.ml.linalg.DenseMatrix
-
- copy() - Method in class org.apache.spark.ml.linalg.DenseVector
-
- copy() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Get a deep copy of the matrix.
- copy() - Method in class org.apache.spark.ml.linalg.SparseMatrix
-
- copy() - Method in class org.apache.spark.ml.linalg.SparseVector
-
- copy() - Method in interface org.apache.spark.ml.linalg.Vector
-
Makes a deep copy of this vector.
- copy(ParamMap) - Method in class org.apache.spark.ml.Model
-
- copy() - Method in class org.apache.spark.ml.param.ParamMap
-
Creates a copy of this param map.
- copy(ParamMap) - Method in interface org.apache.spark.ml.param.Params
-
Creates a copy of this instance with the same UID and some extra params.
- copy(ParamMap) - Method in class org.apache.spark.ml.Pipeline
-
- copy(ParamMap) - Method in class org.apache.spark.ml.PipelineModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.PipelineStage
-
- copy(ParamMap) - Method in class org.apache.spark.ml.Predictor
-
- copy(ParamMap) - Method in class org.apache.spark.ml.recommendation.ALS
-
- copy(ParamMap) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.FMRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.FMRegressor
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.LinearRegression
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- copy(ParamMap) - Method in class org.apache.spark.ml.Transformer
-
- copy(ParamMap) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- copy(ParamMap) - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- copy(ParamMap) - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- copy(ParamMap) - Method in class org.apache.spark.ml.UnaryTransformer
-
- copy(Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
y = x
- copy() - Method in class org.apache.spark.mllib.linalg.DenseMatrix
-
- copy() - Method in class org.apache.spark.mllib.linalg.DenseVector
-
- copy() - Method in interface org.apache.spark.mllib.linalg.Matrix
-
Get a deep copy of the matrix.
- copy() - Method in class org.apache.spark.mllib.linalg.SparseMatrix
-
- copy() - Method in class org.apache.spark.mllib.linalg.SparseVector
-
- copy() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Makes a deep copy of this vector.
- copy() - Method in class org.apache.spark.mllib.random.ExponentialGenerator
-
- copy() - Method in class org.apache.spark.mllib.random.GammaGenerator
-
- copy() - Method in class org.apache.spark.mllib.random.LogNormalGenerator
-
- copy() - Method in class org.apache.spark.mllib.random.PoissonGenerator
-
- copy() - Method in interface org.apache.spark.mllib.random.RandomDataGenerator
-
Returns a copy of the RandomDataGenerator with a new instance of the rng object used in the
class when applicable for non-locking concurrent usage.
- copy() - Method in class org.apache.spark.mllib.random.StandardNormalGenerator
-
- copy() - Method in class org.apache.spark.mllib.random.UniformGenerator
-
- copy() - Method in class org.apache.spark.mllib.random.WeibullGenerator
-
- copy() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
Returns a shallow copy of this instance.
- copy(String, Option<Object>, Option<Object>, Option<Throwable>, Option<String>, Map<String, String>, QueryContext[]) - Method in exception org.apache.spark.sql.AnalysisException
-
- copy() - Method in interface org.apache.spark.sql.Row
-
Make a copy of the current
Row
object.
- copy() - Static method in class org.apache.spark.sql.sources.AlwaysFalse
-
- copy() - Static method in class org.apache.spark.sql.sources.AlwaysTrue
-
- copy() - Method in class org.apache.spark.sql.util.MapperRowCounter
-
- copy() - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- copy() - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- copy() - Method in class org.apache.spark.sql.vectorized.ColumnarMap
-
- copy() - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
Revisit this.
- copy() - Method in class org.apache.spark.util.AccumulatorV2
-
Creates a new copy of this accumulator.
- copy() - Method in class org.apache.spark.util.CollectionAccumulator
-
- copy() - Method in class org.apache.spark.util.DoubleAccumulator
-
- copy() - Method in class org.apache.spark.util.LongAccumulator
-
- copy() - Method in class org.apache.spark.util.StatCounter
-
Clone this StatCounter
- copyAndReset() - Method in class org.apache.spark.sql.util.MapperRowCounter
-
- copyAndReset() - Method in class org.apache.spark.util.AccumulatorV2
-
Creates a new copy of this accumulator, which is zero value.
- copyAndReset() - Method in class org.apache.spark.util.CollectionAccumulator
-
- copyFileStreamNIO(FileChannel, WritableByteChannel, long, long) - Static method in class org.apache.spark.util.Utils
-
- copyStream(InputStream, OutputStream, boolean, boolean) - Static method in class org.apache.spark.util.Utils
-
Copy all data from an InputStream to an OutputStream.
- copyStreamUpTo(InputStream, long) - Static method in class org.apache.spark.util.Utils
-
Copy the first maxSize
bytes of data from the InputStream to an in-memory
buffer, primarily to check for corruption.
- copyValues(T, ParamMap) - Method in interface org.apache.spark.ml.param.Params
-
Copies param values from this instance to another instance for params shared by them.
- cores(int) - Method in class org.apache.spark.resource.ExecutorResourceRequests
-
Specify number of cores per Executor.
- CORES() - Static method in class org.apache.spark.resource.ResourceProfile
-
built-in executor resource: cores
- cores() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
-
- cores() - Method in class org.apache.spark.scheduler.MiscellaneousProcessDetails
-
- CORES_GRANTED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- CORES_PER_EXECUTOR_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- coresGranted() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
-
- coresPerExecutor() - Method in class org.apache.spark.status.api.v1.ApplicationInfo
-
- corr(Dataset<?>, String, String) - Static method in class org.apache.spark.ml.stat.Correlation
-
Compute the correlation matrix for the input Dataset of Vectors using the specified method.
- corr(Dataset<?>, String) - Static method in class org.apache.spark.ml.stat.Correlation
-
Compute the Pearson correlation matrix for the input Dataset of Vectors.
- corr(RDD<Object>, RDD<Object>, String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
-
- corr(RDD<Vector>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Compute the Pearson correlation matrix for the input RDD of Vectors.
- corr(RDD<Vector>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Compute the correlation matrix for the input RDD of Vectors using the specified method.
- corr(RDD<Object>, RDD<Object>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Compute the Pearson correlation for the input RDDs.
- corr(JavaRDD<Double>, JavaRDD<Double>) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of corr()
- corr(RDD<Object>, RDD<Object>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Compute the correlation for the input RDDs using the specified method.
- corr(JavaRDD<Double>, JavaRDD<Double>, String) - Static method in class org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of corr()
- corr(String, String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the correlation of two columns of a DataFrame.
- corr(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the Pearson Correlation Coefficient of two columns of a DataFrame.
- corr(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
- corr(String, String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
- Correlation - Class in org.apache.spark.ml.stat
-
API for correlation functions in MLlib, compatible with DataFrames and Datasets.
- Correlation() - Constructor for class org.apache.spark.ml.stat.Correlation
-
- Correlation - Interface in org.apache.spark.mllib.stat.correlation
-
Trait for correlation algorithms.
- CorrelationNames - Class in org.apache.spark.mllib.stat.correlation
-
Maintains supported and default correlation names.
- CorrelationNames() - Constructor for class org.apache.spark.mllib.stat.correlation.CorrelationNames
-
- Correlations - Class in org.apache.spark.mllib.stat.correlation
-
Delegates computation to the specific correlation object based on the input method name.
- Correlations() - Constructor for class org.apache.spark.mllib.stat.correlation.Correlations
-
- corrMatrix(RDD<Vector>, String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
-
- CORRUPT_MERGED_BLOCK_CHUNKS() - Method in class org.apache.spark.InternalAccumulator.shuffleRead$
-
- CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- corruptedTableNameContextInCatalogError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- corruptedViewQueryOutputColumnsInCatalogError(String, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- corruptedViewReferredTempFunctionsInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- corruptedViewReferredTempViewInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- corruptedViewSQLConfigsInCatalogError(Exception) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- corruptMergedBlockChunks() - Method in class org.apache.spark.status.api.v1.ShufflePushReadMetricDistributions
-
- corruptMergedBlockChunks() - Method in class org.apache.spark.status.api.v1.ShufflePushReadMetrics
-
- cos(Column) - Static method in class org.apache.spark.sql.functions
-
- cos(String) - Static method in class org.apache.spark.sql.functions
-
- cosh(Column) - Static method in class org.apache.spark.sql.functions
-
- cosh(String) - Static method in class org.apache.spark.sql.functions
-
- CosineSilhouette - Class in org.apache.spark.ml.evaluation
-
The algorithm which is implemented in this object, instead, is an efficient and parallel
implementation of the Silhouette using the cosine distance measure.
- CosineSilhouette() - Constructor for class org.apache.spark.ml.evaluation.CosineSilhouette
-
- costSum() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
-
- cot(Column) - Static method in class org.apache.spark.sql.functions
-
- count() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return the number of elements in the RDD.
- count() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
The number of edges in the RDD.
- count() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
The number of vertices in the RDD.
- count() - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
-
- count() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
-
- count(Column, Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- count(Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- count() - Method in class org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Sample size.
- count() - Method in interface org.apache.spark.mllib.stat.MultivariateStatisticalSummary
-
Sample size.
- count() - Method in class org.apache.spark.rdd.RDD
-
Return the number of elements in the RDD.
- Count - Class in org.apache.spark.sql.connector.expressions.aggregate
-
An aggregate function that returns the number of the specific row in a group.
- Count(Expression, boolean) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Count
-
- count() - Method in class org.apache.spark.sql.Dataset
-
Returns the number of rows in the Dataset.
- count(MapFunction<T, Object>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
-
Deprecated.
Count aggregate function.
- count(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
-
Deprecated.
Count aggregate function.
- count(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of items in a group.
- count(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of items in a group.
- count() - Method in class org.apache.spark.sql.KeyValueGroupedDataset
-
Returns a
Dataset
that contains a tuple with each key and the number of items present
for that key.
- count() - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Count the number of rows for each group.
- count(KVStoreView<T>, Function1<T, Object>) - Static method in class org.apache.spark.status.KVUtils
-
Counts the number of elements in the KVStoreView which satisfy a predicate.
- count() - Method in class org.apache.spark.status.RDDPartitionSeq
-
- count() - Method in class org.apache.spark.storage.ReadableChannelFileRegion
-
- count() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by counting each RDD
of this DStream.
- count() - Method in class org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by counting each RDD
of this DStream.
- count() - Method in class org.apache.spark.util.DoubleAccumulator
-
Returns the number of elements added to the accumulator.
- count() - Method in class org.apache.spark.util.LongAccumulator
-
Returns the number of elements added to the accumulator.
- count() - Method in class org.apache.spark.util.StatCounter
-
- count_distinct(Column, Column...) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- count_distinct(Column, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- count_if(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of TRUE
values for the expression.
- count_min_sketch(Column, Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a count-min sketch of a column with the given esp, confidence and seed.
- countApprox(long, double) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApprox(long) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApprox(long, double) - Method in class org.apache.spark.rdd.RDD
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApproxDistinct(double) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinct(int, int) - Method in class org.apache.spark.rdd.RDD
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinct(double) - Method in class org.apache.spark.rdd.RDD
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinctByKey(double, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(int, int, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countAsync() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of count
, which returns a
future for counting the number of elements in this RDD.
- countAsync() - Method in class org.apache.spark.rdd.AsyncRDDActions
-
Returns a future for counting the number of elements in the RDD.
- countByKey() - Method in class org.apache.spark.api.java.JavaPairRDD
-
Count the number of elements for each key, and return the result to the master as a Map.
- countByKey() - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Count the number of elements for each key, collecting the results to a local Map.
- countByKeyApprox(long) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByKeyApprox(long, double) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByKeyApprox(long, double) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByValue() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return the count of each unique value in this RDD as a map of (value, count) pairs.
- countByValue(Ordering<T>) - Method in class org.apache.spark.rdd.RDD
-
Return the count of each unique value in this RDD as a local map of (value, count) pairs.
- countByValue() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValue(int) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValue(int, Ordering<T>) - Method in class org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValueAndWindow(Duration, Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueAndWindow(Duration, Duration, int) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueAndWindow(Duration, Duration, int, Ordering<T>) - Method in class org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueApprox(long, double) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Approximate version of countByValue().
- countByValueApprox(long) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Approximate version of countByValue().
- countByValueApprox(long, double, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
-
Approximate version of countByValue().
- countByValueApproxNotSupportArraysError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- countByWindow(Duration, Duration) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by counting the number
of elements in a window over this DStream.
- countByWindow(Duration, Duration) - Method in class org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by counting the number
of elements in a sliding window over this DStream.
- countDistinct(Column, Column...) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(String, String...) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(Column, Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- COUNTER() - Static method in class org.apache.spark.metrics.sink.StatsdMetricType
-
- CountingWritableChannel - Class in org.apache.spark.storage
-
- CountingWritableChannel(WritableByteChannel) - Constructor for class org.apache.spark.storage.CountingWritableChannel
-
- countMinSketch(String, int, int, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(String, double, double, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(Column, int, int, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(Column, double, double, int) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- CountMinSketch - Class in org.apache.spark.util.sketch
-
A Count-min sketch is a probabilistic data structure used for cardinality estimation using
sub-linear space.
- CountMinSketch() - Constructor for class org.apache.spark.util.sketch.CountMinSketch
-
- CountMinSketch.Version - Enum in org.apache.spark.util.sketch
-
- CountStar - Class in org.apache.spark.sql.connector.expressions.aggregate
-
An aggregate function that returns the number of rows in a group.
- CountStar() - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.CountStar
-
- countTowardsTaskFailures() - Method in class org.apache.spark.ExecutorLostFailure
-
- countTowardsTaskFailures() - Method in class org.apache.spark.FetchFailed
-
Fetch failures lead to a different failure handling path: (1) we don't abort the stage after
4 task failures, instead we immediately go back to the stage which generated the map output,
and regenerate the missing data.
- countTowardsTaskFailures() - Static method in class org.apache.spark.Resubmitted
-
- countTowardsTaskFailures() - Method in class org.apache.spark.TaskCommitDenied
-
If a task failed because its attempt to commit was denied, do not count this failure
towards failing the stage.
- countTowardsTaskFailures() - Method in interface org.apache.spark.TaskFailedReason
-
Whether this task failure should be counted towards the maximum number of times the task is
allowed to fail before the stage is aborted.
- countTowardsTaskFailures() - Method in class org.apache.spark.TaskKilled
-
- countTowardsTaskFailures() - Static method in class org.apache.spark.TaskResultLost
-
- countTowardsTaskFailures() - Static method in class org.apache.spark.UnknownReason
-
- CountVectorizer - Class in org.apache.spark.ml.feature
-
- CountVectorizer(String) - Constructor for class org.apache.spark.ml.feature.CountVectorizer
-
- CountVectorizer() - Constructor for class org.apache.spark.ml.feature.CountVectorizer
-
- CountVectorizerModel - Class in org.apache.spark.ml.feature
-
Converts a text document to a sparse vector of token counts.
- CountVectorizerModel(String, String[]) - Constructor for class org.apache.spark.ml.feature.CountVectorizerModel
-
- CountVectorizerModel(String[]) - Constructor for class org.apache.spark.ml.feature.CountVectorizerModel
-
- CountVectorizerParams - Interface in org.apache.spark.ml.feature
-
- cov() - Method in class org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
- cov(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Calculate the sample covariance of two numerical columns of a DataFrame.
- covar_pop(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the population covariance for two columns.
- covar_pop(String, String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the population covariance for two columns.
- covar_samp(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sample covariance for two columns.
- covar_samp(String, String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sample covariance for two columns.
- covs() - Method in class org.apache.spark.ml.clustering.ExpectationAggregator
-
- cpus() - Method in class org.apache.spark.BarrierTaskContext
-
- CPUS() - Static method in class org.apache.spark.resource.ResourceProfile
-
built-in task resource: cpus
- cpus(int) - Method in class org.apache.spark.resource.TaskResourceRequests
-
Specify number of cpus per Task.
- cpus() - Method in class org.apache.spark.TaskContext
-
CPUs allocated to the task.
- crc32(Column) - Static method in class org.apache.spark.sql.functions
-
Calculates the cyclic redundancy check value (CRC32) of a binary column and
returns the value as a bigint.
- CreatableRelationProvider - Interface in org.apache.spark.sql.sources
-
- create(boolean, boolean, boolean, boolean, int) - Static method in class org.apache.spark.api.java.StorageLevels
-
Create a new StorageLevel object.
- create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int, Function<ResultSet, T>) - Static method in class org.apache.spark.rdd.JdbcRDD
-
Create an RDD that executes a SQL query on a JDBC connection and reads results.
- create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int) - Static method in class org.apache.spark.rdd.JdbcRDD
-
Create an RDD that executes a SQL query on a JDBC connection and reads results.
- create(RDD<T>, Function1<Object, Object>) - Static method in class org.apache.spark.rdd.PartitionPruningRDD
-
Create a PartitionPruningRDD.
- create(String, DataType) - Static method in interface org.apache.spark.sql.connector.catalog.Column
-
- create(String, DataType, boolean) - Static method in interface org.apache.spark.sql.connector.catalog.Column
-
- create(String, DataType, boolean, String, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
-
- create(String, DataType, boolean, String, ColumnDefaultValue, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
-
- create(String, DataType, boolean, String, String, String) - Static method in interface org.apache.spark.sql.connector.catalog.Column
-
- create() - Method in interface org.apache.spark.sql.CreateTableWriter
-
Create a new table from the contents of the data frame.
- create() - Method in class org.apache.spark.sql.DataFrameWriterV2
-
- create(Object...) - Static method in class org.apache.spark.sql.RowFactory
-
Create a
Row
from the given arguments.
- create(long) - Static method in class org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter
with the expected number of insertions and a default expected
false positive probability of 3%.
- create(long, double) - Static method in class org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter
with the expected number of insertions and expected false
positive probability.
- create(long, long) - Static method in class org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter
with given
expectedNumItems
and
numBits
, it will
pick an optimal
numHashFunctions
which can minimize
fpp
for the bloom filter.
- create(int, int, int) - Static method in class org.apache.spark.util.sketch.CountMinSketch
-
- create(double, double, int) - Static method in class org.apache.spark.util.sketch.CountMinSketch
-
Creates a
CountMinSketch
with given relative error (
eps
),
confidence
,
and random
seed
.
- createArrayType(Column) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- createArrayType(DataType) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates an ArrayType by specifying the data type of elements (elementType
).
- createArrayType(DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates an ArrayType by specifying the data type of elements (elementType
) and
whether the array contains null values (containsNull
).
- createArrayWithElementsExceedLimitError(Object) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- createAttrGroupForAttrNames(String, int, boolean, boolean) - Static method in class org.apache.spark.ml.feature.OneHotEncoderCommon
-
Creates an `AttributeGroup` with the required number of `BinaryAttribute`.
- createBatchWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.BatchWrite
-
Creates a writer factory which will be serialized and sent to executors.
- createBatchWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.DeltaBatchWrite
-
- createChunkBlockInfosFromMetaResponse(int, int, int, long, RoaringBitmap[]) - Method in class org.apache.spark.storage.PushBasedFetchHelper
-
This is executed by the task thread when the iterator.next()
is invoked and the iterator
processes a response of type ShuffleBlockFetcherIterator.PushMergedRemoteMetaFetchResult
.
- createColumnarReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.PartitionReaderFactory
-
Returns a columnar partition reader to read data from the given
InputPartition
.
- createColumnarReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReaderFactory
-
- createCombiner() - Method in class org.apache.spark.Aggregator
-
- createCompiledClass(String, File, TestUtils.JavaSourceFromString, Seq<URL>) - Static method in class org.apache.spark.TestUtils
-
Creates a compiled class with the source file.
- createCompiledClass(String, File, String, String, Seq<URL>, Seq<String>, String, Option<String>) - Static method in class org.apache.spark.TestUtils
-
Creates a compiled class with the given name.
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- createConnectionFactory(JDBCOptions) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns a factory for creating connections to the given JDBC URL.
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- createConnectionFactory(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- createContinuousReaderFactory() - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousStream
-
- createCryptoInputStream(InputStream, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
-
Helper method to wrap InputStream
with CryptoInputStream
for decryption.
- createCryptoOutputStream(OutputStream, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
-
Helper method to wrap OutputStream
with CryptoOutputStream
for encryption.
- createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SparkSession
-
Creates a DataFrame
from an RDD of Product (e.g.
- createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SparkSession
-
Creates a DataFrame
from a local Seq of Product.
- createDataFrame(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame
from an
RDD
containing
Row
s using the given schema.
- createDataFrame(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame
from a
JavaRDD
containing
Row
s using the given schema.
- createDataFrame(List<Row>, StructType) - Method in class org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame
from a
java.util.List
containing
Row
s using the given schema.
- createDataFrame(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
-
Applies a schema to an RDD of Java Beans.
- createDataFrame(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
-
Applies a schema to an RDD of Java Beans.
- createDataFrame(List<?>, Class<?>) - Method in class org.apache.spark.sql.SparkSession
-
Applies a schema to a List of Java Beans.
- createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(RDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(JavaRDD<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(List<Row>, StructType) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(RDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(JavaRDD<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataFrame(List<?>, Class<?>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataset(Seq<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
-
Creates a
Dataset
from a local Seq of data of a given type.
- createDataset(RDD<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
-
Creates a
Dataset
from an RDD of a given type.
- createDataset(List<T>, Encoder<T>) - Method in class org.apache.spark.sql.SparkSession
-
Creates a
Dataset
from a
java.util.List
of a given type.
- createDataset(Seq<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataset(RDD<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
-
- createDataset(List<T>, Encoder<T>) - Method in class org.apache.spark.sql.SQLContext
-
- createDayTimeIntervalType(byte, byte) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a DayTimeIntervalType by specifying the start and end fields.
- createDayTimeIntervalType() - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a DayTimeIntervalType with default start and end fields: interval day to second.
- createDecimalType(int, int) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a DecimalType by specifying the precision and scale.
- createDecimalType() - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a DecimalType with default precision and scale, which are 10 and 0.
- createDF(RDD<byte[]>, StructType, SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- createDirectory(File) - Method in interface org.apache.spark.util.SparkFileUtils
-
Create a directory given the abstract pathname
- createDirectory(String, String) - Method in interface org.apache.spark.util.SparkFileUtils
-
Create a directory inside the given parent directory.
- createDirectory(File) - Static method in class org.apache.spark.util.Utils
-
- createDirectory(String, String) - Static method in class org.apache.spark.util.Utils
-
- createDirectory$default$2() - Static method in class org.apache.spark.util.Utils
-
- createEvaluator() - Method in interface org.apache.spark.PartitionEvaluatorFactory
-
Creates a partition evaluator.
- createExternalTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, String) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.SQLContext
-
- createExternalTableWithoutLocationError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- createFailedToGetTokenMessage(String, Throwable) - Static method in class org.apache.spark.util.Utils
-
Returns a string message about delegation token generation failure
- createFuncWithBothIfNotExistsAndReplaceError(SqlBaseParser.CreateFunctionContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- createGlobalTempView(String) - Method in class org.apache.spark.sql.Dataset
-
Creates a global temporary view using the given name.
- createIndex(String, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.index.SupportsIndex
-
Creates an index.
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Build a create index SQL statement.
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- createIndex(String, Identifier, NamedReference[], Map<NamedReference, Map<String, String>>, Map<String, String>) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- createJar(Seq<File>, File, Option<String>, Option<String>) - Static method in class org.apache.spark.TestUtils
-
Create a jar file that contains this set of files.
- createJarWithClasses(Seq<String>, String, Seq<Tuple2<String, String>>, Seq<URL>) - Static method in class org.apache.spark.TestUtils
-
Create a jar that defines classes with the given names.
- createJarWithFiles(Map<String, String>, File) - Static method in class org.apache.spark.TestUtils
-
Create a jar file containing multiple files.
- createKey(SparkConf) - Static method in class org.apache.spark.security.CryptoStreamUtils
-
Creates a new encryption key.
- createKVStore(Option<File>, boolean, SparkConf) - Static method in class org.apache.spark.status.KVUtils
-
- createListeners(SparkConf, ElementTrackingStore) - Method in interface org.apache.spark.status.AppHistoryServerPlugin
-
Creates listeners to replay the event logs.
- createLogForDriver(SparkConf, String, Configuration) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
Create a WriteAheadLog for the driver.
- createLogForReceiver(SparkConf, String, Configuration) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
Create a WriteAheadLog for the receiver.
- createMapOutputWriter(int, long, int) - Method in interface org.apache.spark.shuffle.api.ShuffleExecutorComponents
-
Called once per map task to create a writer that will be responsible for persisting all the
partitioned bytes written by that map task.
- createMapType(DataType, DataType) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a MapType by specifying the data type of keys (keyType
) and values
(keyType
).
- createMapType(DataType, DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a MapType by specifying the data type of keys (keyType
), the data type of
values (keyType
), and whether values contain any null value
(valueContainsNull
).
- createMetrics(long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long) - Static method in class org.apache.spark.status.LiveEntityHelpers
-
- createMetrics(long) - Static method in class org.apache.spark.status.LiveEntityHelpers
-
- createModel(DenseVector<Object>) - Method in interface org.apache.spark.ml.ann.Layer
-
Returns the instance of the layer based on weights provided.
- createNamespace(String[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- createNamespace(String[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsNamespaces
-
Create a namespace in the catalog.
- createOrReplace() - Method in interface org.apache.spark.sql.CreateTableWriter
-
Create a new table or replace an existing table with the contents of the data frame.
- createOrReplace() - Method in class org.apache.spark.sql.DataFrameWriterV2
-
- createOrReplaceGlobalTempView(String) - Method in class org.apache.spark.sql.Dataset
-
Creates or replaces a global temporary view using the given name.
- createOrReplaceTempView(String) - Method in class org.apache.spark.sql.Dataset
-
Creates a local temporary view using the given name.
- createOutputOperationFailureForUI(String) - Static method in class org.apache.spark.streaming.ui.UIUtils
-
- createPartition(InternalRow, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
-
- createPartition(InternalRow, Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.SupportsPartitionManagement
-
Create a partition in table.
- createPartitions(InternalRow[], Map<String, String>[]) - Method in interface org.apache.spark.sql.connector.catalog.SupportsAtomicPartitionManagement
-
Create an array of partitions atomically in table.
- createPersistedViewFromDatasetAPINotAllowedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- createPMMLModelExport(Object) - Static method in class org.apache.spark.mllib.pmml.export.PMMLModelExportFactory
-
Factory object to help creating the necessary PMMLModelExport implementation
taking as input the machine learning model (for example KMeansModel).
- createProxyHandler(Function1<String, Option<String>>) - Static method in class org.apache.spark.ui.JettyUtils
-
Create a handler for proxying request to Workers and Application Drivers
- createProxyLocationHeader(String, HttpServletRequest, URI) - Static method in class org.apache.spark.ui.JettyUtils
-
- createProxyURI(String, String, String, String) - Static method in class org.apache.spark.ui.JettyUtils
-
- createRDDFromArray(JavaSparkContext, byte[][]) - Static method in class org.apache.spark.api.r.RRDD
-
Create an RRDD given a sequence of byte arrays.
- createRDDFromFile(JavaSparkContext, String, int) - Static method in class org.apache.spark.api.r.RRDD
-
Create an RRDD given a temporary file name.
- createReadableChannel(ReadableByteChannel, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
-
Wrap a ReadableByteChannel
for decryption.
- createReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.PartitionReaderFactory
-
Returns a row-based partition reader to read data from the given
InputPartition
.
- createReader(InputPartition) - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReaderFactory
-
- createReaderFactory() - Method in interface org.apache.spark.sql.connector.read.Batch
-
- createReaderFactory() - Method in interface org.apache.spark.sql.connector.read.streaming.MicroBatchStream
-
- createRedirectHandler(String, String, Function1<HttpServletRequest, BoxedUnit>, String, Set<String>) - Static method in class org.apache.spark.ui.JettyUtils
-
Create a handler that always redirects the user to the given path
- createRelation(SQLContext, SaveMode, Map<String, String>, Dataset<Row>) - Method in interface org.apache.spark.sql.sources.CreatableRelationProvider
-
Saves a DataFrame to a destination (using data source-specific parameters)
- createRelation(SQLContext, Map<String, String>) - Method in interface org.apache.spark.sql.sources.RelationProvider
-
Returns a new base relation with the given parameters.
- createRelation(SQLContext, Map<String, String>, StructType) - Method in interface org.apache.spark.sql.sources.SchemaRelationProvider
-
Returns a new base relation with the given parameters and user defined schema.
- createSchedulerBackend(SparkContext, String, TaskScheduler) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
-
Create a scheduler backend for the given SparkContext and scheduler.
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- createSchema(Statement, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Create schema with an optional comment.
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- createSchema(Statement, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- createSecret(SparkConf) - Static method in class org.apache.spark.util.Utils
-
- createServletHandler(String, JettyUtils.ServletParams<T>, SparkConf, String) - Static method in class org.apache.spark.ui.JettyUtils
-
Create a context handler that responds to a request with the given path prefix
- createServletHandler(String, HttpServlet, String) - Static method in class org.apache.spark.ui.JettyUtils
-
Create a context handler that responds to a request with the given path prefix
- createSingleFileMapOutputWriter(int, long) - Method in interface org.apache.spark.shuffle.api.ShuffleExecutorComponents
-
An optional extension for creating a map output writer that can optimize the transfer of a
single partition file, as the entire result of a map task, to the backing store.
- createSink(SQLContext, Map<String, String>, Seq<String>, OutputMode) - Method in interface org.apache.spark.sql.sources.StreamSinkProvider
-
- createSource(SQLContext, String, Option<StructType>, String, Map<String, String>) - Method in interface org.apache.spark.sql.sources.StreamSourceProvider
-
- createSparkContext(String, String, String, String[], Map<Object, Object>, Map<Object, Object>) - Static method in class org.apache.spark.api.r.RRDD
-
- createStaticHandler(String, String) - Static method in class org.apache.spark.ui.JettyUtils
-
Create a handler for serving files from a static directory
- createStream(JavaStreamingContext, String, String, String, String, int, Duration, int, StorageLevel, String, String, String, String, String) - Method in class org.apache.spark.streaming.kinesis.KinesisUtilsPythonHelper
-
- createStreamingSourceNotSpecifySchemaError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- createStreamingWriterFactory(PhysicalWriteInfo) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingWrite
-
Creates a writer factory which will be serialized and sent to executors.
- createStructField(String, String, boolean) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- createStructField(String, DataType, boolean, Metadata) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a StructField by specifying the name (name
), data type (dataType
) and
whether values of this field can be null values (nullable
).
- createStructField(String, DataType, boolean) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a StructField with empty metadata.
- createStructType(Seq<StructField>) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- createStructType(List<StructField>) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a StructType with the given list of StructFields (fields
).
- createStructType(StructField[]) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a StructType with the given StructField array (fields
).
- createTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Creates a table from the given path and returns the corresponding DataFrame.
- createTable(String, String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Creates a table from the given path based on a data source and returns the corresponding
DataFrame.
- createTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
(Scala-specific)
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
(Scala-specific)
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
Create a table based on the dataset in a data source, a schema and a set of options.
- createTable(String, String, StructType, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
(Scala-specific)
Create a table based on the dataset in a data source, a schema and a set of options.
- createTable(String, String, StructType, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
Create a table based on the dataset in a data source, a schema and a set of options.
- createTable(String, String, StructType, String, Map<String, String>) - Method in class org.apache.spark.sql.catalog.Catalog
-
(Scala-specific)
Create a table based on the dataset in a data source, a schema and a set of options.
- createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- createTable(Identifier, Column[], Transform[], Map<String, String>) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- createTable(Identifier, StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
-
Deprecated.
- createTable(Identifier, Column[], Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableCatalog
-
Create a table in the catalog.
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Create the table if the table does not exist.
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- createTable(Statement, String, String, JdbcOptionsInWrite) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- createTableAsSelectWithNonEmptyDirectoryError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- createTableColumnTypesOptionColumnNotFoundInSchemaError(String, StructType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- CreateTableWriter<T> - Interface in org.apache.spark.sql
-
Trait to restrict calls to create and replace operations.
- createTaskScheduler(SparkContext, String) - Method in interface org.apache.spark.scheduler.ExternalClusterManager
-
Create a task scheduler instance for the given SparkContext
- createTempDir() - Method in interface org.apache.spark.util.SparkFileUtils
-
Create a temporary directory inside the java.io.tmpdir
prefixed with spark
.
- createTempDir(String, String) - Method in interface org.apache.spark.util.SparkFileUtils
-
Create a temporary directory inside the given parent directory.
- createTempDir(String, String) - Static method in class org.apache.spark.util.Utils
-
Create a temporary directory inside the given parent directory.
- createTempDir$default$1() - Static method in class org.apache.spark.util.Utils
-
- createTempDir$default$2() - Static method in class org.apache.spark.util.Utils
-
- createTempJsonFile(File, String, JsonAST.JValue) - Static method in class org.apache.spark.TestUtils
-
Creates a temp JSON file that contains the input JSON record.
- createTempScriptWithExpectedOutput(File, String, String) - Static method in class org.apache.spark.TestUtils
-
Creates a temp bash script that prints the given output.
- createTempTableNotSpecifyProviderError(SqlBaseParser.CreateTableContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- createTempView(String) - Method in class org.apache.spark.sql.Dataset
-
Creates a local temporary view using the given name.
- createUnsafe(long, int, int) - Static method in class org.apache.spark.sql.types.Decimal
-
Creates a decimal from unscaled, precision and scale without checking the bounds.
- createView(Identifier, String, String, String[], StructType, String[], String[], String[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.ViewCatalog
-
Create a view in the catalog.
- createViewWithBothIfNotExistsAndReplaceError(SqlBaseParser.CreateViewContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- createWorkspace(int) - Static method in class org.apache.spark.mllib.optimization.NNLS
-
- createWritableChannel(WritableByteChannel, SparkConf, byte[]) - Static method in class org.apache.spark.security.CryptoStreamUtils
-
Wrap a WritableByteChannel
for encryption.
- createWriter(int, long) - Method in interface org.apache.spark.sql.connector.write.DataWriterFactory
-
Returns a data writer to do the actual writing work.
- createWriter(int, long) - Method in interface org.apache.spark.sql.connector.write.DeltaWriterFactory
-
- createWriter(int, long, long) - Method in interface org.apache.spark.sql.connector.write.streaming.StreamingDataWriterFactory
-
Returns a data writer to do the actual writing work.
- createYearMonthIntervalType(byte, byte) - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a YearMonthIntervalType by specifying the start and end fields.
- createYearMonthIntervalType() - Static method in class org.apache.spark.sql.types.DataTypes
-
Creates a YearMonthIntervalType with default start and end fields: interval year to month.
- crossJoin(Dataset<?>) - Method in class org.apache.spark.sql.Dataset
-
Explicit cartesian join with another DataFrame
.
- crosstab(String, String) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Computes a pair-wise frequency table of the given columns.
- CrossValidator - Class in org.apache.spark.ml.tuning
-
K-fold cross validation performs model selection by splitting the dataset into a set of
non-overlapping randomly partitioned folds which are used as separate training and test datasets
e.g., with k=3 folds, K-fold cross validation will generate 3 (training, test) dataset pairs,
each of which uses 2/3 of the data for training and 1/3 for testing.
- CrossValidator(String) - Constructor for class org.apache.spark.ml.tuning.CrossValidator
-
- CrossValidator() - Constructor for class org.apache.spark.ml.tuning.CrossValidator
-
- CrossValidatorModel - Class in org.apache.spark.ml.tuning
-
CrossValidatorModel contains the model with the highest average cross-validation
metric across folds and uses this model to transform input data.
- CrossValidatorModel.CrossValidatorModelWriter - Class in org.apache.spark.ml.tuning
-
Writer for CrossValidatorModel.
- CrossValidatorParams - Interface in org.apache.spark.ml.tuning
-
- CryptoStreamUtils - Class in org.apache.spark.security
-
A util class for manipulating IO encryption and decryption streams.
- CryptoStreamUtils() - Constructor for class org.apache.spark.security.CryptoStreamUtils
-
- CryptoStreamUtils.BaseErrorHandler - Interface in org.apache.spark.security
-
SPARK-25535.
- CryptoStreamUtils.ErrorHandlingReadableChannel - Class in org.apache.spark.security
-
- csc(Column) - Static method in class org.apache.spark.sql.functions
-
- csv(String...) - Method in class org.apache.spark.sql.DataFrameReader
-
Loads CSV files and returns the result as a DataFrame
.
- csv(String) - Method in class org.apache.spark.sql.DataFrameReader
-
Loads a CSV file and returns the result as a DataFrame
.
- csv(Dataset<String>) - Method in class org.apache.spark.sql.DataFrameReader
-
Loads an Dataset[String]
storing CSV rows and returns the result as a DataFrame
.
- csv(Seq<String>) - Method in class org.apache.spark.sql.DataFrameReader
-
Loads CSV files and returns the result as a DataFrame
.
- csv(String) - Method in class org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame
in CSV format at the specified path.
- csv(String) - Method in class org.apache.spark.sql.streaming.DataStreamReader
-
Loads a CSV file stream and returns the result as a DataFrame
.
- cube(Column...) - Method in class org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(String, String...) - Method in class org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- CubeType$() - Constructor for class org.apache.spark.sql.RelationalGroupedDataset.CubeType$
-
- cume_dist() - Static method in class org.apache.spark.sql.functions
-
Window function: returns the cumulative distribution of values within a window partition,
i.e.
- curdate() - Static method in class org.apache.spark.sql.functions
-
Returns the current date at the start of query evaluation as a date column.
- curId() - Static method in class org.apache.spark.sql.Dataset
-
- current_catalog() - Static method in class org.apache.spark.sql.functions
-
Returns the current catalog.
- current_database() - Static method in class org.apache.spark.sql.functions
-
Returns the current database.
- current_date() - Static method in class org.apache.spark.sql.functions
-
Returns the current date at the start of query evaluation as a date column.
- current_schema() - Static method in class org.apache.spark.sql.functions
-
Returns the current schema.
- current_timestamp() - Static method in class org.apache.spark.sql.functions
-
Returns the current timestamp at the start of query evaluation as a timestamp column.
- current_timezone() - Static method in class org.apache.spark.sql.functions
-
Returns the current session local timezone.
- current_user() - Static method in class org.apache.spark.sql.functions
-
Returns the user name of current execution context.
- currentAttemptId() - Method in interface org.apache.spark.SparkStageInfo
-
- currentAttemptId() - Method in class org.apache.spark.SparkStageInfoImpl
-
- currentCatalog() - Method in class org.apache.spark.sql.catalog.Catalog
-
Returns the current catalog in this session.
- currentCatalog() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
Returns the current catalog set.
- currentCatalog() - Method in interface org.apache.spark.sql.connector.catalog.View
-
The current catalog when the view is created.
- currentDatabase() - Method in class org.apache.spark.sql.catalog.Catalog
-
Returns the current database (namespace) in this session.
- currentMetricsValues() - Method in interface org.apache.spark.sql.connector.read.PartitionReader
-
Returns an array of custom task metrics.
- currentMetricsValues() - Method in interface org.apache.spark.sql.connector.write.DataWriter
-
Returns an array of custom task metrics.
- currentNamespace() - Method in interface org.apache.spark.sql.connector.catalog.View
-
The current namespace when the view is created.
- currentResult() - Method in interface org.apache.spark.partial.ApproximateEvaluator
-
- currentRow() - Static method in class org.apache.spark.sql.expressions.Window
-
Value representing the current row.
- currPrefLocs(Partition, RDD<?>) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
- CUSTOM_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- CustomAvgMetric - Class in org.apache.spark.sql.connector.metric
-
Built-in `CustomMetric` that computes average of metric values.
- CustomAvgMetric() - Constructor for class org.apache.spark.sql.connector.metric.CustomAvgMetric
-
- customCollectionClsNotResolvedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- CustomMetric - Interface in org.apache.spark.sql.connector.metric
-
A custom metric.
- customMetrics() - Method in class org.apache.spark.sql.streaming.StateOperatorProgress
-
- CustomSumMetric - Class in org.apache.spark.sql.connector.metric
-
Built-in `CustomMetric` that sums up metric values.
- CustomSumMetric() - Constructor for class org.apache.spark.sql.connector.metric.CustomSumMetric
-
- CustomTaskMetric - Interface in org.apache.spark.sql.connector.metric
-
A custom task metric.
- gain() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- gain() - Method in class org.apache.spark.ml.tree.InternalNode
-
- gain() - Method in class org.apache.spark.mllib.tree.model.InformationGainStats
-
- Gamma$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- gamma1() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma2() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma6() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma7() - Method in class org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- GammaGenerator - Class in org.apache.spark.mllib.random
-
Generates i.i.d.
- GammaGenerator(double, double) - Constructor for class org.apache.spark.mllib.random.GammaGenerator
-
- gammaJavaRDD(JavaSparkContext, double, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.gammaRDD
.
- gammaJavaRDD(JavaSparkContext, double, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaRDD
with the default seed.
- gammaJavaRDD(JavaSparkContext, double, double, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaRDD
with the default number of partitions and the default seed.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.gammaVectorRDD
.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaVectorRDD
with the default seed.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaVectorRDD
with the default number of partitions and the default seed.
- gammaRDD(SparkContext, double, double, long, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD comprised of i.i.d.
samples from the gamma distribution with the input
shape and scale.
- gammaVectorRDD(SparkContext, double, double, long, int, int, long) - Static method in class org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD[Vector] with vectors containing i.i.d.
samples drawn from the
gamma distribution with the input shape and scale.
- gapply(RelationalGroupedDataset, byte[], byte[], Object[], StructType) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
The helper function for gapply() on R side.
- gaps() - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
Indicates whether regex splits on gaps (true) or matches tokens (false).
- GarbageCollectionMetrics - Class in org.apache.spark.metrics
-
- GarbageCollectionMetrics() - Constructor for class org.apache.spark.metrics.GarbageCollectionMetrics
-
- GAUGE() - Static method in class org.apache.spark.metrics.sink.StatsdMetricType
-
- Gaussian$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- GaussianMixture - Class in org.apache.spark.ml.clustering
-
Gaussian Mixture clustering.
- GaussianMixture(String) - Constructor for class org.apache.spark.ml.clustering.GaussianMixture
-
- GaussianMixture() - Constructor for class org.apache.spark.ml.clustering.GaussianMixture
-
- GaussianMixture - Class in org.apache.spark.mllib.clustering
-
This class performs expectation maximization for multivariate Gaussian
Mixture Models (GMMs).
- GaussianMixture() - Constructor for class org.apache.spark.mllib.clustering.GaussianMixture
-
Constructs a default instance.
- GaussianMixtureModel - Class in org.apache.spark.ml.clustering
-
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points
are drawn from each Gaussian i with probability weights(i).
- GaussianMixtureModel - Class in org.apache.spark.mllib.clustering
-
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points
are drawn from each Gaussian i=1..k with probability w(i); mu(i) and sigma(i) are
the respective mean and covariance for each Gaussian distribution i=1..k.
- GaussianMixtureModel(double[], MultivariateGaussian[]) - Constructor for class org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- GaussianMixtureParams - Interface in org.apache.spark.ml.clustering
-
Common params for GaussianMixture and GaussianMixtureModel
- GaussianMixtureSummary - Class in org.apache.spark.ml.clustering
-
Summary of GaussianMixture.
- gaussians() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- gaussians() - Method in class org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- gaussiansDF() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
Retrieve Gaussian distributions as a DataFrame.
- GBTClassificationModel - Class in org.apache.spark.ml.classification
-
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting)
model for classification.
- GBTClassificationModel(String, DecisionTreeRegressionModel[], double[]) - Constructor for class org.apache.spark.ml.classification.GBTClassificationModel
-
Construct a GBTClassificationModel
- GBTClassifier - Class in org.apache.spark.ml.classification
-
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting)
learning algorithm for classification.
- GBTClassifier(String) - Constructor for class org.apache.spark.ml.classification.GBTClassifier
-
- GBTClassifier() - Constructor for class org.apache.spark.ml.classification.GBTClassifier
-
- GBTClassifierParams - Interface in org.apache.spark.ml.tree
-
- GBTParams - Interface in org.apache.spark.ml.tree
-
Parameters for Gradient-Boosted Tree algorithms.
- GBTRegressionModel - Class in org.apache.spark.ml.regression
-
- GBTRegressionModel(String, DecisionTreeRegressionModel[], double[]) - Constructor for class org.apache.spark.ml.regression.GBTRegressionModel
-
Construct a GBTRegressionModel
- GBTRegressor - Class in org.apache.spark.ml.regression
-
- GBTRegressor(String) - Constructor for class org.apache.spark.ml.regression.GBTRegressor
-
- GBTRegressor() - Constructor for class org.apache.spark.ml.regression.GBTRegressor
-
- GBTRegressorParams - Interface in org.apache.spark.ml.tree
-
- GC_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- GC_TIME() - Static method in class org.apache.spark.ui.ToolTips
-
- gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - Static method in class org.apache.spark.ml.linalg.BLAS
-
C := alpha * A * B + beta * C
- gemm(double, Matrix, DenseMatrix, double, double[]) - Static method in class org.apache.spark.ml.linalg.BLAS
-
CValues[0: A.numRows * B.numCols] := alpha * A * B + beta * CValues[0: A.numRows * B.numCols]
- gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
C := alpha * A * B + beta * C
- gemv(double, Matrix, double[], double, double[]) - Static method in class org.apache.spark.ml.linalg.BLAS
-
y[0: A.numRows] := alpha * A * x[0: A.numCols] + beta * y[0: A.numRows]
- gemv(double, Matrix, Vector, double, DenseVector) - Static method in class org.apache.spark.ml.linalg.BLAS
-
y := alpha * A * x + beta * y
- gemv(double, Matrix, Vector, double, DenseVector) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
y := alpha * A * x + beta * y
- GeneralAggregateFunc - Class in org.apache.spark.sql.connector.expressions.aggregate
-
The general implementation of
AggregateFunc
, which contains the upper-cased function
name, the `isDistinct` flag and all the inputs.
- GeneralAggregateFunc(String, boolean, Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.GeneralAggregateFunc
-
- GeneralizedLinearAlgorithm<M extends GeneralizedLinearModel> - Class in org.apache.spark.mllib.regression
-
GeneralizedLinearAlgorithm implements methods to train a Generalized Linear Model (GLM).
- GeneralizedLinearAlgorithm() - Constructor for class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
- GeneralizedLinearModel - Class in org.apache.spark.mllib.regression
-
GeneralizedLinearModel (GLM) represents a model trained using
GeneralizedLinearAlgorithm.
- GeneralizedLinearModel(Vector, double) - Constructor for class org.apache.spark.mllib.regression.GeneralizedLinearModel
-
- GeneralizedLinearRegression - Class in org.apache.spark.ml.regression
-
Fit a Generalized Linear Model
(see
Generalized linear model (Wikipedia))
specified by giving a symbolic description of the linear
predictor (link function) and a description of the error distribution (family).
- GeneralizedLinearRegression(String) - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- GeneralizedLinearRegression() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- GeneralizedLinearRegression.Binomial$ - Class in org.apache.spark.ml.regression
-
Binomial exponential family distribution.
- GeneralizedLinearRegression.CLogLog$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Family$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.FamilyAndLink$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Gamma$ - Class in org.apache.spark.ml.regression
-
Gamma exponential family distribution.
- GeneralizedLinearRegression.Gaussian$ - Class in org.apache.spark.ml.regression
-
Gaussian exponential family distribution.
- GeneralizedLinearRegression.Identity$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Inverse$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Link$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Log$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Logit$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Poisson$ - Class in org.apache.spark.ml.regression
-
Poisson exponential family distribution.
- GeneralizedLinearRegression.Probit$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Sqrt$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegression.Tweedie$ - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegressionBase - Interface in org.apache.spark.ml.regression
-
Params for Generalized Linear Regression.
- GeneralizedLinearRegressionModel - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegressionSummary - Class in org.apache.spark.ml.regression
-
- GeneralizedLinearRegressionTrainingSummary - Class in org.apache.spark.ml.regression
-
- GeneralMLWritable - Interface in org.apache.spark.ml.util
-
Trait for classes that provide GeneralMLWriter
.
- GeneralMLWriter - Class in org.apache.spark.ml.util
-
A ML Writer which delegates based on the requested format.
- GeneralMLWriter(PipelineStage) - Constructor for class org.apache.spark.ml.util.GeneralMLWriter
-
- GeneralScalarExpression - Class in org.apache.spark.sql.connector.expressions
-
The general representation of SQL scalar expressions, which contains the upper-cased
expression name and all the children expressions.
- GeneralScalarExpression(String, Expression[]) - Constructor for class org.apache.spark.sql.connector.expressions.GeneralScalarExpression
-
- generateAssociationRules(double) - Method in class org.apache.spark.mllib.fpm.FPGrowthModel
-
Generates association rules for the Item
s in freqItemsets
.
- generateKMeansRDD(SparkContext, int, int, int, double, int) - Static method in class org.apache.spark.mllib.util.KMeansDataGenerator
-
Generate an RDD containing test data for KMeans.
- generateLinearInput(double, double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
-
For compatibility, the generated data without specifying the mean and variance
will have zero mean and variance of (1.0/3.0) since the original output range is
[-1, 1] with uniform distribution, and the variance of uniform distribution
is (b - a)^2^ / 12 which will be (1.0/3.0)
- generateLinearInput(double, double[], double[], double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
-
- generateLinearInput(double, double[], double[], double[], int, int, double, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
-
- generateLinearInputAsList(double, double[], int, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
-
Return a Java List of synthetic data randomly generated according to a multi
collinear model.
- generateLinearRDD(SparkContext, int, int, double, int, double) - Static method in class org.apache.spark.mllib.util.LinearDataGenerator
-
Generate an RDD containing sample data for Linear Regression models - including Ridge, Lasso,
and unregularized variants.
- generateLogisticRDD(SparkContext, int, int, double, int, double) - Static method in class org.apache.spark.mllib.util.LogisticRegressionDataGenerator
-
Generate an RDD containing test data for LogisticRegression.
- generateRandomEdges(int, int, int, long) - Static method in class org.apache.spark.graphx.util.GraphGenerators
-
- generateRolledOverFileSuffix() - Method in interface org.apache.spark.util.logging.RollingPolicy
-
Get the desired name of the rollover file
- generationExpression() - Method in interface org.apache.spark.sql.connector.catalog.Column
-
Returns the generation expression of this table column.
- generatorNotExpectedError(FunctionIdentifier, String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- generatorOutsideSelectError(LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- geq(Object) - Method in class org.apache.spark.sql.Column
-
Greater than or equal to an expression.
- get(Object) - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- get() - Method in class org.apache.spark.api.java.Optional
-
- get() - Static method in class org.apache.spark.BarrierTaskContext
-
:: Experimental ::
Returns the currently active BarrierTaskContext.
- get() - Method in interface org.apache.spark.FutureAction
-
Blocks and returns the result of this job.
- get(Param<T>) - Method in class org.apache.spark.ml.param.ParamMap
-
Optionally returns the value associated with a param.
- get(Param<T>) - Method in interface org.apache.spark.ml.param.Params
-
Optionally returns the user-supplied value of a param.
- get(String) - Method in class org.apache.spark.SparkConf
-
Get a parameter; throws a NoSuchElementException if it's not set
- get(String, String) - Method in class org.apache.spark.SparkConf
-
Get a parameter, falling back to a default if not set
- get() - Static method in class org.apache.spark.SparkEnv
-
Returns the SparkEnv.
- get(String) - Static method in class org.apache.spark.SparkFiles
-
Get the absolute path of a file added through SparkContext.addFile()
.
- get() - Method in interface org.apache.spark.sql.connector.read.PartitionReader
-
Return the current record.
- get(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns element of array at given (0-based) index.
- get(String) - Static method in class org.apache.spark.sql.jdbc.JdbcDialects
-
Fetch the JdbcDialect class corresponding to a given database url.
- get() - Method in class org.apache.spark.sql.Observation
-
(Scala-specific) Get the observed metrics.
- get(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i.
- get(String) - Method in class org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- get(String, String) - Method in class org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- get() - Method in interface org.apache.spark.sql.streaming.GroupState
-
Get the state value if it exists, or throw NoSuchElementException.
- get(UUID) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns the query if there is an active query with the given id, or null.
- get(String) - Method in class org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns the query if there is an active query with the given id, or null.
- get(Object) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- get(int, DataType) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- get() - Method in class org.apache.spark.streaming.State
-
Get the state if it exists, otherwise it will throw java.util.NoSuchElementException
.
- get() - Static method in class org.apache.spark.TaskContext
-
Return the currently active TaskContext.
- get(long) - Static method in class org.apache.spark.util.AccumulatorContext
-
- get_json_object(Column, String) - Static method in class org.apache.spark.sql.functions
-
Extracts json object from a json string based on json path specified, and returns json string
of the extracted json object.
- getAbsolutePathFromExecutable(String) - Static method in class org.apache.spark.TestUtils
-
Get the absolute path from the executable.
- getAcceptanceResults(RDD<Tuple2<K, V>>, boolean, Map<K, Object>, Option<Map<K, Object>>, long) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Count the number of items instantly accepted and generate the waitlist for each stratum.
- getAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
int64 accumulator_id = 2;
- getAccumulatorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
int64 accumulator_id = 2;
- getAccumulatorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
-
int64 accumulator_id = 2;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdates(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdates(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getAccumulatorUpdatesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- getActive() - Static method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Get the currently active context, if there is one.
- getActiveJobIds() - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns an array containing the ids of all active jobs.
- getActiveJobIds() - Method in class org.apache.spark.SparkStatusTracker
-
Returns an array containing the ids of all active jobs.
- getActiveOrCreate(Function0<StreamingContext>) - Static method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Either return the "active" StreamingContext (that is, started but not stopped), or create a
new StreamingContext that is
- getActiveOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Either get the currently active StreamingContext (that is, started but not stopped),
OR recreate a StreamingContext from checkpoint data in the given path.
- getActiveSession() - Static method in class org.apache.spark.sql.SparkSession
-
Returns the active SparkSession for the current thread, returned by the builder.
- getActiveStageIds() - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns an array containing the ids of all active stages.
- getActiveStageIds() - Method in class org.apache.spark.SparkStatusTracker
-
Returns an array containing the ids of all active stages.
- getActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 active_tasks = 9;
- getActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 active_tasks = 9;
- getActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 active_tasks = 9;
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getAddColumnQuery(String, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getAddColumnQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional string address = 1;
- getAddress() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional string address = 1;
- getAddress() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional string address = 1;
- getAddressBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional string address = 1;
- getAddressBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional string address = 1;
- getAddressBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional string address = 1;
- getAddresses(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- getAddresses(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
repeated string addresses = 2;
- getAddresses(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
repeated string addresses = 2;
- getAddressesBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- getAddressesBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
repeated string addresses = 2;
- getAddressesBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
repeated string addresses = 2;
- getAddressesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- getAddressesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
repeated string addresses = 2;
- getAddressesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
repeated string addresses = 2;
- getAddressesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- getAddressesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
repeated string addresses = 2;
- getAddressesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
repeated string addresses = 2;
- getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 add_time = 20;
- getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 add_time = 20;
- getAddTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 add_time = 20;
- getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int64 add_time = 5;
- getAddTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
int64 add_time = 5;
- getAddTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
int64 add_time = 5;
- getAggregationDepth() - Method in interface org.apache.spark.ml.param.shared.HasAggregationDepth
-
- getAlgo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getAll() - Method in class org.apache.spark.SparkConf
-
Get all parameters as a list of pairs
- getAll() - Method in class org.apache.spark.sql.RuntimeConfig
-
Returns all properties set in this conf.
- getAllClusterConfigs(SparkConf) - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
-
- getAllConfs() - Method in class org.apache.spark.sql.SQLContext
-
Return all the configuration properties that have been set (i.e.
- getAllPools() - Method in class org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return pools for fair scheduler
- GetAllReceiverInfo - Class in org.apache.spark.streaming.scheduler
-
- GetAllReceiverInfo() - Constructor for class org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- getAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_removals_time_ms = 6;
- getAllRemovalsTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 all_removals_time_ms = 6;
- getAllRemovalsTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 all_removals_time_ms = 6;
- getAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_updates_time_ms = 4;
- getAllUpdatesTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 all_updates_time_ms = 4;
- getAllUpdatesTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 all_updates_time_ms = 4;
- getAllWithPrefix(String) - Method in class org.apache.spark.SparkConf
-
Get all parameters that start with prefix
- getAlpha() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getAlpha() - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for getDocConcentration
- getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
int64 amount = 2;
- getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
int64 amount = 2;
- getAmount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
int64 amount = 2;
- getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
double amount = 2;
- getAmount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
double amount = 2;
- getAmount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
-
double amount = 2;
- getAnyValAs(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i.
- getAppId() - Method in interface org.apache.spark.launcher.SparkAppHandle
-
Returns the application ID, or null
if not yet known.
- getAppId() - Method in class org.apache.spark.SparkConf
-
Returns the Spark application id, valid in the Driver after TaskScheduler registration and
from the start in the Executor.
- getApplicationInfo(String) - Method in interface org.apache.spark.status.api.v1.UIRoot
-
- getApplicationInfoList() - Method in interface org.apache.spark.status.api.v1.UIRoot
-
- getAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string app_spark_version = 8;
- getAppSparkVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string app_spark_version = 8;
- getAppSparkVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string app_spark_version = 8;
- getAppSparkVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string app_spark_version = 8;
- getAppSparkVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string app_spark_version = 8;
- getAppSparkVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string app_spark_version = 8;
- getArray(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getArray(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the array type value for rowId
.
- getAs(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i.
- getAs(String) - Method in interface org.apache.spark.sql.Row
-
Returns the value of a given fieldName.
- getAsJava() - Method in class org.apache.spark.sql.Observation
-
(Java-specific) Get the observed metrics.
- getAssociationRulesFromFP(Dataset<?>, String, String, double, Map<T, Object>, long, ClassTag<T>) - Static method in class org.apache.spark.ml.fpm.AssociationRules
-
Computes the association rules with confidence above minConfidence.
- getAsymmetricAlpha() - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for getAsymmetricDocConcentration
- getAsymmetricDocConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 attempt = 3;
- getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int32 attempt = 3;
- getAttempt() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int32 attempt = 3;
- getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 attempt = 3;
- getAttempt() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int32 attempt = 3;
- getAttempt() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int32 attempt = 3;
- getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string attempt_id = 1;
- getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string attempt_id = 1;
- getAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string attempt_id = 1;
- getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 attempt_id = 3;
- getAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 attempt_id = 3;
- getAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 attempt_id = 3;
- getAttemptIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string attempt_id = 1;
- getAttemptIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string attempt_id = 1;
- getAttemptIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string attempt_id = 1;
- getAttempts(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttempts(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttempts(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttemptsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- getAttr(String) - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its name.
- getAttr(int) - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its index.
- getAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
Deprecated.
- getAttributes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
Deprecated.
- getAttributesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getAttributesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getAttributesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> attributes = 27;
- getAttributesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> attributes = 27;
- getAttributesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> attributes = 27;
- getAttributesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> attributes = 27;
- getAttributesOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> attributes = 27;
- getAttributesOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> attributes = 27;
- getAttributesOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> attributes = 27;
- getAttributesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> attributes = 27;
- getAttributesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> attributes = 27;
- getAttributesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> attributes = 27;
- getAvroField(String, int) - Method in class org.apache.spark.sql.avro.AvroUtils.AvroSchemaHelper
-
Get the Avro field corresponding to the provided Catalyst field name/position, if any.
- getAvroSchema() - Method in class org.apache.spark.SparkConf
-
Gets all the avro schemas in the configuration used in the generic Avro record serializer
- getBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool barrier = 4;
- getBarrier() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
bool barrier = 4;
- getBarrier() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
bool barrier = 4;
- getBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_duration = 6;
- getBatchDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
int64 batch_duration = 6;
- getBatchDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
int64 batch_duration = 6;
- getBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_id = 5;
- getBatchId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
int64 batch_id = 5;
- getBatchId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
int64 batch_id = 5;
- getBatchingTimeout(SparkConf) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
How long we will wait for the wrappedLog in the BatchedWriteAheadLog to write the records
before we fail the write attempt to unblock receivers.
- getBernoulliSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Return the per partition sampling function used for sampling without replacement.
- getBeta() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getBeta() - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for getTopicConcentration
- getBin(int) - Method in class org.apache.spark.sql.util.NumericHistogram
-
Returns a particular histogram bin.
- getBinary() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
-
- getBinary() - Method in class org.apache.spark.ml.feature.HashingTF
-
- getBinary(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getBinary(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the binary type value for rowId
.
- getbit(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns the value of the bit (0 or 1) at the specified position.
- getBlacklistedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 blacklisted_in_stages = 25;
- getBlacklistedInStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 blacklisted_in_stages = 25;
- getBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string block_name = 1;
- getBlockName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
optional string block_name = 1;
- getBlockName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
optional string block_name = 1;
- getBlockNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string block_name = 1;
- getBlockNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
optional string block_name = 1;
- getBlockNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
optional string block_name = 1;
- getBlockSize() - Method in interface org.apache.spark.ml.param.shared.HasBlockSize
-
- GetBlockStatus(BlockId, boolean) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
-
- GetBlockStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetBlockStatus$
-
- getBoolean(String, boolean) - Method in class org.apache.spark.SparkConf
-
Get a parameter as a boolean, falling back to a default if not set
- getBoolean(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive boolean.
- getBoolean(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Boolean.
- getBoolean(String, boolean) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
Returns the boolean value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getBoolean(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the boolean type value for rowId
.
- getBooleanArray(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Boolean array.
- getBooleans(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets boolean type values from [rowId, rowId + count)
.
- getBootstrap() - Method in interface org.apache.spark.ml.tree.RandomForestParams
-
- getBucketLength() - Method in interface org.apache.spark.ml.feature.BucketedRandomProjectionLSHParams
-
- getBuilder() - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
-
- getBuilder() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
-
- getBuilder() - Method in interface org.apache.spark.storage.memory.ValuesHolder
-
Note: After this method is called, the ValuesHolder is invalid, we can't store data and
get estimate size again.
- getByte(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive byte.
- getByte(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getByte(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the byte type value for rowId
.
- getBytes(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets byte type values from [rowId, rowId + count)
.
- getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_read = 18;
- getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double bytes_read = 18;
- getBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double bytes_read = 18;
- getBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- getBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double bytes_read = 1;
- getBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double bytes_read = 1;
- getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 bytes_read = 1;
- getBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
int64 bytes_read = 1;
- getBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricsOrBuilder
-
int64 bytes_read = 1;
- getBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- getBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double bytes_read = 1;
- getBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double bytes_read = 1;
- getBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- getBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double bytes_read = 1;
- getBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double bytes_read = 1;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_written = 20;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double bytes_written = 20;
- getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double bytes_written = 20;
- getBytesWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- getBytesWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double bytes_written = 1;
- getBytesWritten(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double bytes_written = 1;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 bytes_written = 1;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
int64 bytes_written = 1;
- getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricsOrBuilder
-
int64 bytes_written = 1;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 bytes_written = 1;
- getBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
int64 bytes_written = 1;
- getBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
-
int64 bytes_written = 1;
- getBytesWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- getBytesWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double bytes_written = 1;
- getBytesWrittenCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double bytes_written = 1;
- getBytesWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- getBytesWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double bytes_written = 1;
- getBytesWrittenList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double bytes_written = 1;
- getCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool cached = 3;
- getCached() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
bool cached = 3;
- getCached() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
bool cached = 3;
- getCachedBlockManagerId(BlockManagerId) - Static method in class org.apache.spark.storage.BlockManagerId
-
- getCachedMetadata(String) - Static method in class org.apache.spark.rdd.HadoopRDD
-
The three methods below are helpers for accessing the local map, a property of the SparkEnv of
the local process.
- getCacheNodeIds() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string callsite = 5;
- getCallsite() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
optional string callsite = 5;
- getCallsite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
optional string callsite = 5;
- getCallSite(Function1<String, Object>) - Static method in class org.apache.spark.util.Utils
-
When called inside a class in the spark package, returns the name of the user code class
(outside the spark package) that called into Spark, as well as which Spark method they called.
- getCallsiteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string callsite = 5;
- getCallsiteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
optional string callsite = 5;
- getCallsiteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
optional string callsite = 5;
- getCaseSensitive() - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- getCatalystType(int, String, int, MetadataBuilder) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Get the custom datatype mapping for the given jdbc meta information.
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getCategoricalCols() - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- getCategoricalFeatures(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
-
Examine a schema to identify categorical (Binary and Nominal) features.
- getCategoricalFeaturesInfo() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getCensorCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
-
- getCheckpointDir() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- getCheckpointDir() - Method in class org.apache.spark.SparkContext
-
- getCheckpointFile() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Gets the name of the file to which this RDD was checkpointed
- getCheckpointFile() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
- getCheckpointFile() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
- getCheckpointFile() - Method in class org.apache.spark.rdd.RDD
-
Gets the name of the directory to which this RDD was checkpointed.
- getCheckpointFiles() - Method in class org.apache.spark.graphx.Graph
-
Gets the name of the files to which this Graph was checkpointed.
- getCheckpointFiles() - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- getCheckpointFiles() - Method in class org.apache.spark.ml.clustering.DistributedLDAModel
-
If using checkpointing and LDA.keepLastCheckpoint
is set to true, then there may be
saved checkpoint files.
- getCheckpointInterval() - Method in interface org.apache.spark.ml.param.shared.HasCheckpointInterval
-
- getCheckpointInterval() - Method in class org.apache.spark.mllib.clustering.LDA
-
Period (in iterations) between checkpoints.
- getCheckpointInterval() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getChild(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getChild(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
- getChildClusters(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClusters(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClusters(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildClustersOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- getChildNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getChildNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- getClassifier() - Method in interface org.apache.spark.ml.classification.OneVsRestParams
-
- getClasspathEntries(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntries(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntries(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getClasspathEntriesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- getCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getCluster() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getClusterBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getClusterConfig(SparkConf, String) - Static method in class org.apache.spark.kafka010.KafkaTokenSparkConf
-
- getClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getClusterOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- getColdStartStrategy() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
-
- getCollectSubModels() - Method in interface org.apache.spark.ml.param.shared.HasCollectSubModels
-
- getColumnName(Seq<Object>, StructType) - Static method in class org.apache.spark.sql.util.SchemaUtils
-
Gets the name of the column in the given position.
- getCombOp() - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Returns the function used combine results returned by seqOp from different partitions.
- getComment() - Method in class org.apache.spark.sql.types.StructField
-
Return the comment of this StructField.
- getCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 commit_time_ms = 7;
- getCommitTimeMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 commit_time_ms = 7;
- getCommitTimeMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 commit_time_ms = 7;
- getCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
bool completed = 7;
- getCompleted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
bool completed = 7;
- getCompleted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
bool completed = 7;
- getCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 completed_tasks = 11;
- getCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 completed_tasks = 11;
- getCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 completed_tasks = 11;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 completion_time = 5;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional int64 completion_time = 5;
- getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional int64 completion_time = 5;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional int64 completion_time = 9;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional int64 completion_time = 9;
- getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional int64 completion_time = 9;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 completion_time = 12;
- getCompletionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional int64 completion_time = 12;
- getCompletionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional int64 completion_time = 12;
- getConf() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Return a copy of this JavaSparkContext's configuration.
- getConf() - Method in interface org.apache.spark.input.Configurable
-
- getConf() - Method in class org.apache.spark.rdd.HadoopRDD
-
- getConf() - Method in class org.apache.spark.rdd.NewHadoopRDD
-
- getConf() - Method in class org.apache.spark.SparkContext
-
Return a copy of this SparkContext's configuration.
- getConf(String) - Method in class org.apache.spark.sql.SQLContext
-
Return the value of Spark SQL configuration property for the given key.
- getConf(String, String) - Method in class org.apache.spark.sql.SQLContext
-
Return the value of Spark SQL configuration property for the given key.
- getConfiguration() - Method in class org.apache.spark.input.PortableDataStream
-
- getConfiguredLocalDirs(SparkConf) - Static method in class org.apache.spark.util.Utils
-
Return the configured local directories where Spark can write files.
- getConnection() - Method in interface org.apache.spark.rdd.JdbcRDD.ConnectionFactory
-
- getConnection(Driver, Map<String, String>) - Method in class org.apache.spark.sql.jdbc.JdbcConnectionProvider
-
Opens connection to the database.
- getContextOrSparkClassLoader() - Method in interface org.apache.spark.util.SparkClassUtils
-
- getContextOrSparkClassLoader() - Static method in class org.apache.spark.util.Utils
-
- getConvergenceTol() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Return the largest change in log-likelihood at which convergence is
considered to have occurred.
- getCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_granted = 3;
- getCoresGranted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional int32 cores_granted = 3;
- getCoresGranted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional int32 cores_granted = 3;
- getCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_per_executor = 5;
- getCoresPerExecutor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional int32 cores_per_executor = 5;
- getCoresPerExecutor() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional int32 cores_per_executor = 5;
- getCorrelationFromName(String) - Static method in class org.apache.spark.mllib.stat.correlation.Correlations
-
- getCorruptMergedBlockChunks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double corrupt_merged_block_chunks = 1;
- getCorruptMergedBlockChunksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double corrupt_merged_block_chunks = 1;
- getCount() - Method in class org.apache.spark.storage.CountingWritableChannel
-
- getCurrentDefaultValue() - Method in class org.apache.spark.sql.types.StructField
-
Return the current default value of this StructField.
- getCurrentProcessingTimeMs() - Method in interface org.apache.spark.sql.streaming.GroupState
-
Get the current processing time as milliseconds in epoch time.
- getCurrentUserGroups(SparkConf, String) - Static method in class org.apache.spark.util.Utils
-
- getCurrentUserName() - Static method in class org.apache.spark.util.Utils
-
Returns the current user name.
- getCurrentWatermarkMs() - Method in interface org.apache.spark.sql.streaming.GroupState
-
Get the current event time watermark as milliseconds in epoch time.
- getCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
Deprecated.
- getCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
Deprecated.
- getCustomMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
Deprecated.
- getCustomMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- getCustomMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getCustomMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
map<string, int64> custom_metrics = 12;
- getCustomMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
map<string, int64> custom_metrics = 12;
- getData(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the image data
- getDatabase(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Get the database (namespace) with the specified name (can be qualified with catalog).
- getDataDistribution(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistribution(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistribution(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDataDistributionOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- getDate(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.sql.Date.
- getDecimal(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of decimal type as java.math.BigDecimal.
- getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getDecimal(int, int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the decimal type value for rowId
.
- getDefault(Param<T>) - Method in interface org.apache.spark.ml.param.Params
-
Gets the default value of a parameter.
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- getDefaultInstance() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- getDefaultInstanceForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- getDefaultPropertiesFile(Map<String, String>) - Static method in class org.apache.spark.util.Utils
-
Return the path of the default Spark properties file.
- getDefaultReadLimit() - Method in interface org.apache.spark.sql.connector.read.streaming.SupportsAdmissionControl
-
Returns the read limits potentially passed to the data source through options when creating
the data source.
- getDefaultSession() - Static method in class org.apache.spark.sql.SparkSession
-
Returns the default SparkSession that is returned by the builder.
- getDegree() - Method in class org.apache.spark.ml.feature.PolynomialExpansion
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getDeleteColumnQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getDeleteColumnQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getDenseSizeInBytes() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Gets the size of the dense representation of this `Matrix`.
- getDependencies() - Method in class org.apache.spark.rdd.CoGroupedRDD
-
- getDependencies() - Method in class org.apache.spark.rdd.ShuffledRDD
-
- getDependencies() - Method in class org.apache.spark.rdd.UnionRDD
-
- getDeprecatedConfig(String, Map<String, String>) - Static method in class org.apache.spark.SparkConf
-
Looks for available deprecated keys for the given config option, and return the first
value available.
- getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string desc = 3;
- getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
optional string desc = 3;
- getDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
optional string desc = 3;
- getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string desc = 3;
- getDesc() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
optional string desc = 3;
- getDesc() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
optional string desc = 3;
- getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string desc = 3;
- getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
optional string desc = 3;
- getDescBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
optional string desc = 3;
- getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string desc = 3;
- getDescBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
optional string desc = 3;
- getDescBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
optional string desc = 3;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string description = 3;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string description = 3;
- getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string description = 3;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
optional string description = 1;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
optional string description = 1;
- getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
optional string description = 1;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string description = 1;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string description = 1;
- getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string description = 1;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string description = 3;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string description = 3;
- getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string description = 3;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string description = 40;
- getDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string description = 40;
- getDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string description = 40;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string description = 3;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string description = 3;
- getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string description = 3;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
optional string description = 1;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
optional string description = 1;
- getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
optional string description = 1;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string description = 1;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string description = 1;
- getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string description = 1;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string description = 3;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string description = 3;
- getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string description = 3;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string description = 40;
- getDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string description = 40;
- getDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string description = 40;
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getDescriptor() - Static method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- getDescriptor() - Static method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- getDescriptorForType() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- getDescriptorForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- getDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool deserialized = 7;
- getDeserialized() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
bool deserialized = 7;
- getDeserialized() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
bool deserialized = 7;
- getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string details = 4;
- getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string details = 4;
- getDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string details = 4;
- getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string details = 41;
- getDetails() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string details = 41;
- getDetails() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string details = 41;
- getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string details = 4;
- getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string details = 4;
- getDetailsBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string details = 4;
- getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string details = 41;
- getDetailsBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string details = 41;
- getDetailsBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string details = 41;
- getDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string discoveryScript = 3;
- getDiscoveryScript() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string discoveryScript = 3;
- getDiscoveryScript() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string discoveryScript = 3;
- getDiscoveryScriptBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string discoveryScript = 3;
- getDiscoveryScriptBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string discoveryScript = 3;
- getDiscoveryScriptBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string discoveryScript = 3;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double disk_bytes_spilled = 17;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double disk_bytes_spilled = 17;
- getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double disk_bytes_spilled = 17;
- getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 disk_bytes_spilled = 14;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 disk_bytes_spilled = 14;
- getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 disk_bytes_spilled = 14;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 disk_bytes_spilled = 22;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 disk_bytes_spilled = 22;
- getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 disk_bytes_spilled = 22;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 disk_bytes_spilled = 24;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 disk_bytes_spilled = 24;
- getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 disk_bytes_spilled = 24;
- getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 disk_bytes_spilled = 9;
- getDiskBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 disk_bytes_spilled = 9;
- getDiskBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 disk_bytes_spilled = 9;
- getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 15;
- getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double disk_bytes_spilled = 14;
- getDiskBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double disk_bytes_spilled = 14;
- getDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 disk_size = 9;
- getDiskSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
int64 disk_size = 9;
- getDiskSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
int64 disk_size = 9;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 disk_used = 6;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 disk_used = 6;
- getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 disk_used = 6;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 disk_used = 4;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
int64 disk_used = 4;
- getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
int64 disk_used = 4;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 disk_used = 4;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
int64 disk_used = 4;
- getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
int64 disk_used = 4;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 disk_used = 7;
- getDiskUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
int64 disk_used = 7;
- getDiskUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
int64 disk_used = 7;
- getDistanceMeasure() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getDistanceMeasure() - Method in class org.apache.spark.ml.evaluation.ClusteringMetrics
-
- getDistanceMeasure() - Method in interface org.apache.spark.ml.param.shared.HasDistanceMeasure
-
- getDistanceMeasure() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
The distance suite used by the algorithm.
- getDistanceMeasure() - Method in class org.apache.spark.mllib.clustering.KMeans
-
The distance suite used by the algorithm.
- getDistributions() - Method in class org.apache.spark.status.LiveRDD
-
- getDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getDocConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- getDouble(String, double) - Method in class org.apache.spark.SparkConf
-
Get a parameter as a double, falling back to a default if not ste
- getDouble(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive double.
- getDouble(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Double.
- getDouble(String, double) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
Returns the double value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getDouble(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getDouble(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the double type value for rowId
.
- getDoubleArray(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Double array.
- getDoubles(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets double type values from [rowId, rowId + count)
.
- getDriverAttributes() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
Get the attributes on driver.
- getDriverLogUrls() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
Get the URLs for the driver logs.
- getDropLast() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
-
- getDstCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
-
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 duration = 5;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
int64 duration = 5;
- getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
int64 duration = 5;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double duration = 5;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double duration = 5;
- getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double duration = 5;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 duration = 7;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional int64 duration = 7;
- getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional int64 duration = 7;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 duration = 7;
- getDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 duration = 7;
- getDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 duration = 7;
- getDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- getDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double duration = 2;
- getDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double duration = 2;
- getDuration(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
-
- getDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- getDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double duration = 2;
- getDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double duration = 2;
- getDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- getDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double duration = 2;
- getDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double duration = 2;
- getDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
Deprecated.
- getDurationMs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
Deprecated.
- getDurationMsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getDurationMsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getDurationMsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, int64> duration_ms = 7;
- getDurationMsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, int64> duration_ms = 7;
- getDurationMsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, int64> duration_ms = 7;
- getDurationMsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, int64> duration_ms = 7;
- getDurationMsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, int64> duration_ms = 7;
- getDurationMsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, int64> duration_ms = 7;
- getDurationMsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, int64> duration_ms = 7;
- getDurationMsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, int64> duration_ms = 7;
- getDurationMsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, int64> duration_ms = 7;
- getDurationMsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, int64> duration_ms = 7;
- getDynamicAllocationInitialExecutors(SparkConf) - Static method in class org.apache.spark.util.Utils
-
Return the initial number of executors for dynamic allocation.
- getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- getElasticNetParam() - Method in interface org.apache.spark.ml.param.shared.HasElasticNetParam
-
- getEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string end_offset = 3;
- getEndOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string end_offset = 3;
- getEndOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string end_offset = 3;
- getEndOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string end_offset = 3;
- getEndOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string end_offset = 3;
- getEndOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string end_offset = 3;
- getEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 end_time = 3;
- getEndTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
int64 end_time = 3;
- getEndTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
int64 end_time = 3;
- getEndTimeEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional int64 end_timestamp = 7;
- getEndTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional int64 end_timestamp = 7;
- getEndTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional int64 end_timestamp = 7;
- getEps() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getEpsilon() - Method in interface org.apache.spark.ml.regression.LinearRegressionParams
-
- getEpsilon() - Method in class org.apache.spark.mllib.clustering.KMeans
-
The distance threshold within which we've consider centers to have converged.
- getError() - Method in interface org.apache.spark.launcher.SparkAppHandle
-
If the application failed due to an error, return the underlying error.
- getErrorClass() - Method in exception org.apache.spark.SparkException
-
- getErrorClass() - Method in interface org.apache.spark.SparkThrowable
-
- getErrorClass() - Method in exception org.apache.spark.sql.AnalysisException
-
- getErrorClass() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
-
- getErrorMessage(String, Map<String, String>) - Method in class org.apache.spark.ErrorClassesJsonReader
-
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string error_message = 10;
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string error_message = 10;
- getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string error_message = 10;
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string error_message = 14;
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string error_message = 14;
- getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string error_message = 14;
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string error_message = 14;
- getErrorMessage() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string error_message = 14;
- getErrorMessage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string error_message = 10;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string error_message = 10;
- getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string error_message = 10;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string error_message = 14;
- getErrorMessageBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string error_message = 14;
- getEstimator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
-
- getEstimatorParamMaps() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
-
- getEvaluator() - Method in interface org.apache.spark.ml.tuning.ValidatorParams
-
- getEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
Deprecated.
- getEventTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
Deprecated.
- getEventTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getEventTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getEventTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> event_time = 8;
- getEventTimeMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> event_time = 8;
- getEventTimeMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> event_time = 8;
- getEventTimeMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> event_time = 8;
- getEventTimeOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> event_time = 8;
- getEventTimeOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> event_time = 8;
- getEventTimeOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> event_time = 8;
- getEventTimeOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> event_time = 8;
- getEventTimeOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> event_time = 8;
- getEventTimeOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> event_time = 8;
- getException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string exception = 5;
- getException() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string exception = 5;
- getException() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string exception = 5;
- getExceptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string exception = 5;
- getExceptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string exception = 5;
- getExceptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string exception = 5;
- getExcludedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
repeated int64 excluded_in_stages = 31;
- getExcludedInStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
repeated int64 excluded_in_stages = 31;
- getExecutionContext() - Method in interface org.apache.spark.ml.param.shared.HasParallelism
-
Create a new execution context with a thread-pool that has a maximum number of threads
set to the value of parallelism
.
- getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
int64 execution_id = 1;
- getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
int64 execution_id = 1;
- getExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
int64 execution_id = 1;
- getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 execution_id = 1;
- getExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
int64 execution_id = 1;
- getExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
int64 execution_id = 1;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_cpu_time = 9;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double executor_cpu_time = 9;
- getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double executor_cpu_time = 9;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_cpu_time = 17;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 executor_cpu_time = 17;
- getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 executor_cpu_time = 17;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_cpu_time = 19;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 executor_cpu_time = 19;
- getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 executor_cpu_time = 19;
- getExecutorCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_cpu_time = 4;
- getExecutorCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 executor_cpu_time = 4;
- getExecutorCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 executor_cpu_time = 4;
- getExecutorCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_cpu_time = 6;
- getExecutorCpuTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_cpu_time = 6;
- getExecutorDecommissionState(String) - Method in interface org.apache.spark.scheduler.TaskScheduler
-
If an executor is decommissioned, return its corresponding decommission info
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_cpu_time = 7;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double executor_deserialize_cpu_time = 7;
- getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double executor_deserialize_cpu_time = 7;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_cpu_time = 15;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 executor_deserialize_cpu_time = 15;
- getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 executor_deserialize_cpu_time = 15;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_cpu_time = 17;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 executor_deserialize_cpu_time = 17;
- getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 executor_deserialize_cpu_time = 17;
- getExecutorDeserializeCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_cpu_time = 2;
- getExecutorDeserializeCpuTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 executor_deserialize_cpu_time = 2;
- getExecutorDeserializeCpuTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 executor_deserialize_cpu_time = 2;
- getExecutorDeserializeCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeCpuTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_cpu_time = 4;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_time = 6;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double executor_deserialize_time = 6;
- getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double executor_deserialize_time = 6;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_time = 14;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 executor_deserialize_time = 14;
- getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 executor_deserialize_time = 14;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_time = 16;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 executor_deserialize_time = 16;
- getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 executor_deserialize_time = 16;
- getExecutorDeserializeTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_time = 1;
- getExecutorDeserializeTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 executor_deserialize_time = 1;
- getExecutorDeserializeTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 executor_deserialize_time = 1;
- getExecutorDeserializeTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_deserialize_time = 3;
- getExecutorDeserializeTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_deserialize_time = 3;
- GetExecutorEndpointRef(String) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef
-
- GetExecutorEndpointRef$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef$
-
- getExecutorEnv() - Method in class org.apache.spark.SparkConf
-
Get all executor environment variables set on this SparkConf
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
optional string executor_id = 3;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
optional string executor_id = 3;
- getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
optional string executor_id = 3;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string executor_id = 2;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string executor_id = 2;
- getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string executor_id = 2;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string executor_id = 8;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string executor_id = 8;
- getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string executor_id = 8;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string executor_id = 8;
- getExecutorId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string executor_id = 8;
- getExecutorId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
optional string executor_id = 3;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
optional string executor_id = 3;
- getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
optional string executor_id = 3;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string executor_id = 2;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string executor_id = 2;
- getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string executor_id = 2;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string executor_id = 8;
- getExecutorIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string executor_id = 8;
- getExecutorInfos() - Method in class org.apache.spark.SparkStatusTracker
-
Returns information of all known executors, including host, port, cacheSize, numRunningTasks
and memory metrics.
- getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
Deprecated.
- getExecutorLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
Deprecated.
- getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
Deprecated.
- getExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
Deprecated.
- getExecutorLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
Deprecated.
- getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getExecutorLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> executor_logs = 23;
- getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- getExecutorLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getExecutorLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
map<string, string> executor_logs = 16;
- getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> executor_logs = 23;
- getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> executor_logs = 23;
- getExecutorLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> executor_logs = 23;
- getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
map<string, string> executor_logs = 16;
- getExecutorLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
map<string, string> executor_logs = 16;
- getExecutorLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, string> executor_logs = 23;
- getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
map<string, string> executor_logs = 16;
- getExecutorLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
map<string, string> executor_logs = 16;
- GetExecutorLossReason(String) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason
-
- GetExecutorLossReason$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason$
-
- getExecutorMemoryStatus() - Method in class org.apache.spark.SparkContext
-
Return a map from the block manager to the max memory available for caching and the remaining
memory available for caching.
- getExecutorMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsDistributionsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- getExecutorMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- getExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
Deprecated.
- getExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
Deprecated.
- getExecutorResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
Deprecated.
- getExecutorResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- getExecutorResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getExecutorResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrDefault(String, StoreTypes.ExecutorResourceRequest) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorResourceRequest> executor_resources = 2;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_run_time = 8;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double executor_run_time = 8;
- getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double executor_run_time = 8;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_run_time = 16;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 executor_run_time = 16;
- getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 executor_run_time = 16;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_run_time = 18;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 executor_run_time = 18;
- getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 executor_run_time = 18;
- getExecutorRunTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- getExecutorRunTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_run_time = 5;
- getExecutorRunTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_run_time = 5;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_run_time = 3;
- getExecutorRunTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 executor_run_time = 3;
- getExecutorRunTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 executor_run_time = 3;
- getExecutorRunTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- getExecutorRunTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_run_time = 5;
- getExecutorRunTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_run_time = 5;
- getExecutorRunTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- getExecutorRunTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double executor_run_time = 5;
- getExecutorRunTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double executor_run_time = 5;
- getExecutors(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- getExecutors(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
repeated string executors = 5;
- getExecutors(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
repeated string executors = 5;
- getExecutorsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- getExecutorsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
repeated string executors = 5;
- getExecutorsBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
repeated string executors = 5;
- getExecutorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- getExecutorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
repeated string executors = 5;
- getExecutorsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
repeated string executors = 5;
- getExecutorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- getExecutorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
repeated string executors = 5;
- getExecutorsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
repeated string executors = 5;
- getExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
Deprecated.
- getExecutorSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
Deprecated.
- getExecutorSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getExecutorSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getExecutorSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrDefault(String, StoreTypes.ExecutorStageSummary) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getExecutorSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ExecutorStageSummary> executor_summary = 46;
- getFactorSize() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
-
- getFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- getFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double failed_tasks = 3;
- getFailedTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double failed_tasks = 3;
- getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 failed_tasks = 2;
- getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int32 failed_tasks = 2;
- getFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int32 failed_tasks = 2;
- getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 failed_tasks = 10;
- getFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 failed_tasks = 10;
- getFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 failed_tasks = 10;
- getFailedTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- getFailedTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double failed_tasks = 3;
- getFailedTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double failed_tasks = 3;
- getFailedTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- getFailedTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double failed_tasks = 3;
- getFailedTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double failed_tasks = 3;
- getFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string failure_reason = 13;
- getFailureReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string failure_reason = 13;
- getFailureReason() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string failure_reason = 13;
- getFailureReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string failure_reason = 13;
- getFailureReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string failure_reason = 13;
- getFailureReasonBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string failure_reason = 13;
- getFamily() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
- getFamily() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getFdr() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getFeatureIndex() - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
-
- getFeatureIndicesFromNames(StructField, String[]) - Static method in class org.apache.spark.ml.util.MetadataUtils
-
Takes a Vector column and a list of feature names, and returns the corresponding list of
feature indices in the column, in order.
- getFeatures() - Method in class org.apache.spark.ml.feature.LabeledPoint
-
- getFeatures() - Method in class org.apache.spark.mllib.regression.LabeledPoint
-
- getFeaturesAndLabels(RFormulaModel, Dataset<?>) - Static method in class org.apache.spark.ml.r.RWrapperUtils
-
Get the feature names and original labels from the schema
of DataFrame transformed by RFormulaModel.
- getFeaturesCol() - Method in interface org.apache.spark.ml.param.shared.HasFeaturesCol
-
- getFeatureSubsetStrategy() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
-
- getFeatureType() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
- getFetchWaitTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- getFetchWaitTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double fetch_wait_time = 5;
- getFetchWaitTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double fetch_wait_time = 5;
- getFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 fetch_wait_time = 3;
- getFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 fetch_wait_time = 3;
- getFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 fetch_wait_time = 3;
- getFetchWaitTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- getFetchWaitTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double fetch_wait_time = 5;
- getFetchWaitTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double fetch_wait_time = 5;
- getFetchWaitTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- getFetchWaitTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double fetch_wait_time = 5;
- getFetchWaitTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double fetch_wait_time = 5;
- getField(String) - Method in class org.apache.spark.sql.Column
-
An expression that gets a field by name in a StructType
.
- getFileLength(File, SparkConf) - Static method in class org.apache.spark.util.Utils
-
Return the file length, if the file is compressed it returns the uncompressed file length.
- getFileSegmentLocations(String, long, long, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
Get the locations of the HDFS blocks containing the given file segment.
- getFileSystemForPath(Path, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
- getFinalStorageLevel() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getFinalValue() - Method in class org.apache.spark.partial.PartialResult
-
Blocking method to wait for and return the final value.
- getFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 first_task_launched_time = 11;
- getFirstTaskLaunchedTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional int64 first_task_launched_time = 11;
- getFirstTaskLaunchedTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional int64 first_task_launched_time = 11;
- getFitIntercept() - Method in interface org.apache.spark.ml.param.shared.HasFitIntercept
-
- getFitLinear() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
-
- getFloat(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive float.
- getFloat(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getFloat(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the float type value for rowId
.
- getFloats(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets float type values from [rowId, rowId + count)
.
- getFoldCol() - Method in interface org.apache.spark.ml.tuning.CrossValidatorParams
-
- getForceIndexLabel() - Method in interface org.apache.spark.ml.feature.RFormulaBase
-
- getFormattedClassName(Object) - Static method in class org.apache.spark.util.Utils
-
Return the class name of the given object, removing all dollar signs
- getFormattedDuration(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
-
- getFormattedSubmissionTime(JobData) - Static method in class org.apache.spark.ui.jobs.JobDataUtil
-
- getFormula() - Method in interface org.apache.spark.ml.feature.RFormulaBase
-
- getFpr() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 from_id = 1;
- getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
int32 from_id = 1;
- getFromId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdgeOrBuilder
-
int32 from_id = 1;
- getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 from_id = 1;
- getFromId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
int64 from_id = 1;
- getFromId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdgeOrBuilder
-
int64 from_id = 1;
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Return the DB-specific quoted and fully qualified table name
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getFullyQualifiedQuotedTableName(Identifier) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getFunction(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Get the function with the specified name.
- getFunction(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Get the function with the specified name in the specified database under the Hive Metastore.
- getFwe() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getGaps() - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double getting_result_time = 13;
- getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double getting_result_time = 13;
- getGettingResultTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double getting_result_time = 13;
- getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 getting_result_time = 18;
- getGettingResultTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int64 getting_result_time = 18;
- getGettingResultTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int64 getting_result_time = 18;
- getGettingResultTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- getGettingResultTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double getting_result_time = 10;
- getGettingResultTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double getting_result_time = 10;
- getGettingResultTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- getGettingResultTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double getting_result_time = 10;
- getGettingResultTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double getting_result_time = 10;
- getGettingResultTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- getGettingResultTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double getting_result_time = 10;
- getGettingResultTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double getting_result_time = 10;
- getGroups(String) - Method in interface org.apache.spark.security.GroupMappingServiceProvider
-
Get the groups the user belongs to.
- getHadoopFileSystem(URI, Configuration) - Static method in class org.apache.spark.util.Utils
-
Return a Hadoop FileSystem with the scheme encoded in the given path.
- getHadoopFileSystem(String, Configuration) - Static method in class org.apache.spark.util.Utils
-
Return a Hadoop FileSystem with the scheme encoded in the given path.
- getHadoopProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHadoopPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- getHandleInvalid() - Method in interface org.apache.spark.ml.param.shared.HasHandleInvalid
-
- getHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool has_metrics = 15;
- getHasMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
bool has_metrics = 15;
- getHasMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
bool has_metrics = 15;
- getHeapHistogram() - Static method in class org.apache.spark.util.Utils
-
Return a heap dump.
- getHeight(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the height of the image
- getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string host = 9;
- getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string host = 9;
- getHost() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string host = 9;
- getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string host = 9;
- getHost() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string host = 9;
- getHost() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string host = 9;
- getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string host = 9;
- getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string host = 9;
- getHostBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string host = 9;
- getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string host = 9;
- getHostBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string host = 9;
- getHostBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string host = 9;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string host_port = 2;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string host_port = 2;
- getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string host_port = 2;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string host_port = 2;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
optional string host_port = 2;
- getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
optional string host_port = 2;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string host_port = 3;
- getHostPort() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string host_port = 3;
- getHostPort() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string host_port = 3;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string host_port = 2;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string host_port = 2;
- getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string host_port = 2;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string host_port = 2;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
optional string host_port = 2;
- getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
optional string host_port = 2;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string host_port = 3;
- getHostPortBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string host_port = 3;
- getHostPortBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string host_port = 3;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
int64 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional string id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
optional string id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
optional string id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
int32 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
int32 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
int32 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
int32 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
int64 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
int64 id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
int64 id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string id = 2;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string id = 2;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string id = 2;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string id = 1;
- getId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string id = 1;
- getId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional string id = 1;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string id = 1;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
optional string id = 1;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
optional string id = 1;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string id = 2;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string id = 2;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string id = 2;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string id = 1;
- getIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string id = 1;
- getIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string id = 1;
- getImplicitPrefs() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getImpurity() - Method in interface org.apache.spark.ml.tree.HasVarianceImpurity
-
- getImpurity() - Method in interface org.apache.spark.ml.tree.TreeClassifierParams
-
- getImpurity() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getIncomingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIncomingEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 index = 2;
- getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int32 index = 2;
- getIndex() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int32 index = 2;
- getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 index = 2;
- getIndex() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int32 index = 2;
- getIndex() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int32 index = 2;
- getIndices() - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapperOrBuilder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- getInitializationMode() - Method in class org.apache.spark.mllib.clustering.KMeans
-
The initialization algorithm.
- getInitializationSteps() - Method in class org.apache.spark.mllib.clustering.KMeans
-
Number of steps for the k-means|| initialization mode
- getInitialModel() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Return the user supplied initial GMM, if supplied
- getInitialTargetExecutorNumber(SparkConf, int) - Static method in class org.apache.spark.scheduler.cluster.SchedulerBackendUtils
-
Getting the initial target number of executors depends on whether dynamic allocation is
enabled.
- getInitialWeights() - Method in interface org.apache.spark.ml.classification.MultilayerPerceptronParams
-
- getInitMode() - Method in interface org.apache.spark.ml.clustering.KMeansParams
-
- getInitMode() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
-
- getInitStd() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
-
- getInitSteps() - Method in interface org.apache.spark.ml.clustering.KMeansParams
-
- getInOutCols() - Method in interface org.apache.spark.ml.feature.ImputerParams
-
Returns the input and output column names corresponding in pair.
- getInOutCols() - Method in interface org.apache.spark.ml.feature.OneHotEncoderBase
-
Returns the input and output column names corresponding in pair.
- getInOutCols() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
-
Returns the input and output column names corresponding in pair.
- getInputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- getInputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_bytes = 6;
- getInputBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_bytes = 6;
- getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_bytes = 5;
- getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 input_bytes = 5;
- getInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 input_bytes = 5;
- getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_bytes = 24;
- getInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 input_bytes = 24;
- getInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 input_bytes = 24;
- getInputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- getInputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_bytes = 6;
- getInputBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_bytes = 6;
- getInputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- getInputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_bytes = 6;
- getInputBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_bytes = 6;
- getInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_bytes_read = 26;
- getInputBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 input_bytes_read = 26;
- getInputBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 input_bytes_read = 26;
- getInputCol() - Method in interface org.apache.spark.ml.param.shared.HasInputCol
-
- getInputCols() - Method in interface org.apache.spark.ml.param.shared.HasInputCols
-
- getInputFilePath() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
-
Returns the holding file name or empty string if it is unknown.
- getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- getInputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- getInputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_records = 7;
- getInputRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_records = 7;
- getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_records = 6;
- getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 input_records = 6;
- getInputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 input_records = 6;
- getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_records = 25;
- getInputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 input_records = 25;
- getInputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 input_records = 25;
- getInputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- getInputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_records = 7;
- getInputRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_records = 7;
- getInputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- getInputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double input_records = 7;
- getInputRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double input_records = 7;
- getInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_records_read = 27;
- getInputRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 input_records_read = 27;
- getInputRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 input_records_read = 27;
- getInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double input_rows_per_second = 6;
- getInputRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
double input_rows_per_second = 6;
- getInputRowsPerSecond() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
double input_rows_per_second = 6;
- getInputStream(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
- getInstant(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.time.Instant.
- getInt(String, int) - Method in class org.apache.spark.SparkConf
-
Get a parameter as an integer, falling back to a default if not set
- getInt(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive int.
- getInt(String, int) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
Returns the integer value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getInt(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getInt(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the int type value for rowId
.
- getIntermediateStorageLevel() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getInterval(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the calendar interval type value for rowId
.
- getInts(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets int type values from [rowId, rowId + count)
.
- getInverse() - Method in class org.apache.spark.ml.feature.DCT
-
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_active = 3;
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
bool is_active = 3;
- getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
bool is_active = 3;
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
bool is_active = 3;
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
bool is_active = 3;
- getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
bool is_active = 3;
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
bool is_active = 4;
- getIsActive() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
bool is_active = 4;
- getIsActive() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
bool is_active = 4;
- getIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_blacklisted = 18;
- getIsBlacklisted() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
bool is_blacklisted = 18;
- getIsBlacklisted() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
bool is_blacklisted = 18;
- getIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_blacklisted_for_stage = 15;
- getIsBlacklistedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
bool is_blacklisted_for_stage = 15;
- getIsBlacklistedForStage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
bool is_blacklisted_for_stage = 15;
- getIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_excluded = 30;
- getIsExcluded() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
bool is_excluded = 30;
- getIsExcluded() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
bool is_excluded = 30;
- getIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_excluded_for_stage = 17;
- getIsExcludedForStage() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
bool is_excluded_for_stage = 17;
- getIsExcludedForStage() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
bool is_excluded_for_stage = 17;
- getIsExperiment() - Method in class org.apache.spark.mllib.stat.test.BinarySample
-
- getIsotonic() - Method in interface org.apache.spark.ml.regression.IsotonicRegressionBase
-
- getIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
bool is_shuffle_push_enabled = 63;
- getIsShufflePushEnabled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
bool is_shuffle_push_enabled = 63;
- getIsShufflePushEnabled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
bool is_shuffle_push_enabled = 63;
- getItem(Object) - Method in class org.apache.spark.sql.Column
-
An expression that gets an item at position ordinal
out of an array,
or gets a value by key key
in a MapType
.
- getItemCol() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
-
- getItemsCol() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
-
- getIteratorSize(Iterator<?>) - Static method in class org.apache.spark.util.Utils
-
Counts the number of elements of an iterator.
- getIteratorZipWithIndex(Iterator<T>, long) - Static method in class org.apache.spark.util.Utils
-
Generate a zipWithIndex iterator, avoid index value overflowing problem
in scala's zipWithIndex
- getIvyProperties() - Static method in class org.apache.spark.util.DependencyUtils
-
- getJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_home = 2;
- getJavaHome() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string java_home = 2;
- getJavaHome() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string java_home = 2;
- getJavaHomeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_home = 2;
- getJavaHomeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string java_home = 2;
- getJavaHomeBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string java_home = 2;
- getJavaMap(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of array type as a java.util.Map
.
- getJavaSparkContext(SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- getJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_version = 1;
- getJavaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string java_version = 1;
- getJavaVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string java_version = 1;
- getJavaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_version = 1;
- getJavaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string java_version = 1;
- getJavaVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string java_version = 1;
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns the SQL builder for the SELECT statement.
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getJdbcSQLQueryBuilder(JDBCOptions) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getJDBCType(DataType) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getJDBCType(DataType) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Retrieve the jdbc / sql type for a given datatype.
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getJDBCType(DataType) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string job_group = 7;
- getJobGroup() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string job_group = 7;
- getJobGroup() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string job_group = 7;
- getJobGroupBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string job_group = 7;
- getJobGroupBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string job_group = 7;
- getJobGroupBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string job_group = 7;
- getJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
All IDs are int64 for extendability, even when they are currently int32 in Spark.
- getJobId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
All IDs are int64 for extendability, even when they are currently int32 in Spark.
- getJobId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
All IDs are int64 for extendability, even when they are currently int32 in Spark.
- getJobIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- getJobIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
repeated int64 job_ids = 2;
- getJobIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
repeated int64 job_ids = 2;
- getJobIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- getJobIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
repeated int64 job_ids = 2;
- getJobIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
repeated int64 job_ids = 2;
- getJobIdsForGroup(String) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
-
Return a list of all known jobs in a particular job group.
- getJobIdsForGroup(String) - Method in class org.apache.spark.SparkStatusTracker
-
Return a list of all known jobs in a particular job group.
- getJobIdsForTag(String) - Method in class org.apache.spark.SparkStatusTracker
-
Return a list of all known jobs with a particular tag.
- getJobIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- getJobIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
repeated int64 job_ids = 2;
- getJobIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
repeated int64 job_ids = 2;
- getJobInfo(int) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns job information, or null
if the job info could not be found or was garbage collected.
- getJobInfo(int) - Method in class org.apache.spark.SparkStatusTracker
-
Returns job information, or None
if the job info could not be found or was garbage collected.
- getJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
Deprecated.
- getJobs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
Deprecated.
- getJobsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getJobsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getJobsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrDefault(long, StoreTypes.JobExecutionStatus) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
Deprecated.
- getJobsValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
Deprecated.
- getJobsValueMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrDefault(long, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrDefault(long, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrDefault(long, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobsValueOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.JobExecutionStatus> jobs = 11;
- getJobTags() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Get the tags that are currently set to be assigned to all the jobs started by this thread.
- getJobTags(JavaSparkContext) - Static method in class org.apache.spark.api.r.RUtils
-
- getJobTags() - Method in class org.apache.spark.SparkContext
-
Get the tags that are currently set to be assigned to all the jobs started by this thread.
- getJobTags(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- getJobTags(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated string job_tags = 21;
- getJobTags(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated string job_tags = 21;
- getJobTagsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- getJobTagsBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated string job_tags = 21;
- getJobTagsBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated string job_tags = 21;
- getJobTagsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- getJobTagsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated string job_tags = 21;
- getJobTagsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated string job_tags = 21;
- getJobTagsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- getJobTagsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated string job_tags = 21;
- getJobTagsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated string job_tags = 21;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double jvm_gc_time = 11;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double jvm_gc_time = 11;
- getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double jvm_gc_time = 11;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 jvm_gc_time = 19;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 jvm_gc_time = 19;
- getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 jvm_gc_time = 19;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 jvm_gc_time = 21;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 jvm_gc_time = 21;
- getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 jvm_gc_time = 21;
- getJvmGcTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- getJvmGcTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double jvm_gc_time = 8;
- getJvmGcTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double jvm_gc_time = 8;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 jvm_gc_time = 6;
- getJvmGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 jvm_gc_time = 6;
- getJvmGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 jvm_gc_time = 6;
- getJvmGcTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- getJvmGcTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double jvm_gc_time = 8;
- getJvmGcTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double jvm_gc_time = 8;
- getJvmGcTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- getJvmGcTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double jvm_gc_time = 8;
- getJvmGcTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double jvm_gc_time = 8;
- getK() - Method in interface org.apache.spark.ml.clustering.BisectingKMeansParams
-
- getK() - Method in interface org.apache.spark.ml.clustering.GaussianMixtureParams
-
- getK() - Method in interface org.apache.spark.ml.clustering.KMeansParams
-
- getK() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getK() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
-
- getK() - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- getK() - Method in interface org.apache.spark.ml.feature.PCAParams
-
- getK() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the desired number of leaf clusters.
- getK() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Return the number of Gaussians in the mixture model
- getK() - Method in class org.apache.spark.mllib.clustering.KMeans
-
Number of clusters to create (k).
- getK() - Method in class org.apache.spark.mllib.clustering.LDA
-
Number of topics to infer, i.e., the number of soft cluster centers.
- getKappa() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Learning rate: exponential decay rate
- getKeepLastCheckpoint() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getKeepLastCheckpoint() - Method in class org.apache.spark.mllib.clustering.EMLDAOptimizer
-
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
- getKeytabJaasParams(String, String, String) - Static method in class org.apache.spark.kafka010.KafkaTokenUtil
-
- getKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- getKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double killed_tasks = 5;
- getKilledTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double killed_tasks = 5;
- getKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 killed_tasks = 4;
- getKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int32 killed_tasks = 4;
- getKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int32 killed_tasks = 4;
- getKilledTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- getKilledTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double killed_tasks = 5;
- getKilledTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double killed_tasks = 5;
- getKilledTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- getKilledTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double killed_tasks = 5;
- getKilledTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double killed_tasks = 5;
- getKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
Deprecated.
- getKilledTasksSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
Deprecated.
- getKilledTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getKilledTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getKilledTasksSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrDefault(String, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<string, int32> killed_tasks_summary = 48;
- getKilledTasksSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<string, int32> killed_tasks_summary = 48;
- getKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
Deprecated.
- getKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
Deprecated.
- getKillTasksSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
Deprecated.
- getKillTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- getKillTasksSummaryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getKillTasksSummaryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrDefault(String, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrDefault(String, int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
map<string, int32> kill_tasks_summary = 20;
- getKillTasksSummaryOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
map<string, int32> kill_tasks_summary = 20;
- getKrb5LoginModuleName() - Static method in class org.apache.spark.util.SecurityUtils
-
Krb5LoginModule package varies in different JVMs.
- getLabel() - Method in class org.apache.spark.ml.feature.LabeledPoint
-
- getLabel() - Method in class org.apache.spark.mllib.regression.LabeledPoint
-
- getLabelCol() - Method in interface org.apache.spark.ml.param.shared.HasLabelCol
-
- getLabels() - Method in class org.apache.spark.ml.feature.IndexToString
-
- getLabelType() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
- getLambda() - Method in class org.apache.spark.mllib.classification.NaiveBayes
-
Get the smoothing parameter.
- getLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 last_updated = 4;
- getLastUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
int64 last_updated = 4;
- getLastUpdated() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
int64 last_updated = 4;
- getLastUpdatedEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string latest_offset = 4;
- getLatestOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string latest_offset = 4;
- getLatestOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string latest_offset = 4;
- getLatestOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string latest_offset = 4;
- getLatestOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string latest_offset = 4;
- getLatestOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string latest_offset = 4;
- getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 launch_time = 5;
- getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int64 launch_time = 5;
- getLaunchTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int64 launch_time = 5;
- getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 launch_time = 5;
- getLaunchTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 launch_time = 5;
- getLaunchTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 launch_time = 5;
- getLayers() - Method in interface org.apache.spark.ml.classification.MultilayerPerceptronParams
-
- getLDAModel(double[]) - Method in interface org.apache.spark.mllib.clustering.LDAOptimizer
-
- getLeafCol() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getLeafField(String) - Method in interface org.apache.spark.ml.tree.DecisionTreeModel
-
- getLeafField(String) - Method in interface org.apache.spark.ml.tree.TreeEnsembleModel
-
- getLearningDecay() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getLearningOffset() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getLearningRate() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getLeastGroupHash(String) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
Gets the least element of the list associated with key in groupHash
The returned PartitionGroup is the least loaded of all groups that represent the machine "key"
- getLength() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
-
Returns the length of the block being read, or -1 if it is unknown.
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getLimitClause(Integer) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns the LIMIT clause for the SELECT statement
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getLimitClause(Integer) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getLink() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getLinkPower() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getLinkPredictionCol() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getList(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of array type as java.util.List
.
- getLocalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_blocks_fetched = 2;
- getLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 local_blocks_fetched = 2;
- getLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 local_blocks_fetched = 2;
- getLocalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double local_blocks_fetched = 4;
- getLocalBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double local_blocks_fetched = 4;
- getLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_bytes_read = 6;
- getLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 local_bytes_read = 6;
- getLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 local_bytes_read = 6;
- getLocalDate(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.time.LocalDate.
- getLocalDir(SparkConf) - Static method in class org.apache.spark.util.Utils
-
Get the path of a temporary directory.
- getLocale() - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- getLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
Deprecated.
- getLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
Deprecated.
- getLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
Deprecated.
- getLocalityCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- getLocalityCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getLocalityCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
map<string, int64> locality = 3;
- getLocalityMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
map<string, int64> locality = 3;
- getLocalityMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
map<string, int64> locality = 3;
- getLocalityMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
map<string, int64> locality = 3;
- getLocalityOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
map<string, int64> locality = 3;
- getLocalityOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
map<string, int64> locality = 3;
- getLocalityOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
map<string, int64> locality = 3;
- getLocalityOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
map<string, int64> locality = 3;
- getLocalityOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
map<string, int64> locality = 3;
- getLocalityOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder
-
map<string, int64> locality = 3;
- getLocalMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_blocks_fetched = 4;
- getLocalMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_bytes_read = 8;
- getLocalMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 local_merged_bytes_read = 8;
- getLocalMergedBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 local_merged_bytes_read = 8;
- getLocalMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_bytes_read = 8;
- getLocalMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double local_merged_chunks_fetched = 6;
- getLocalMergedChunksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double local_merged_chunks_fetched = 6;
- getLocalProperty(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Get a local property set in this thread, or null if it is missing.
- getLocalProperty(String) - Method in class org.apache.spark.BarrierTaskContext
-
- getLocalProperty(String) - Method in class org.apache.spark.SparkContext
-
Get a local property set in this thread, or null if it is missing.
- getLocalProperty(String) - Method in class org.apache.spark.TaskContext
-
Get a local property set upstream in the driver, or null if it is missing.
- getLocalUserJarsForShell(SparkConf) - Static method in class org.apache.spark.util.Utils
-
Return the local jar files which will be added to REPL's classpath.
- GetLocations(BlockId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocations
-
- GetLocations$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocations$
-
- GetLocationsAndStatus(BlockId, String) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus
-
- GetLocationsAndStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus$
-
- GetLocationsMultipleBlockIds(BlockId[]) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds
-
- GetLocationsMultipleBlockIds$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds$
-
- getLong(String, long) - Method in class org.apache.spark.SparkConf
-
Get a parameter as a long, falling back to a default if not set
- getLong(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive long.
- getLong(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Long.
- getLong(String, long) - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
Returns the long value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getLong(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getLong(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the long type value for rowId
.
- getLongArray(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Long array.
- getLongs(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets long type values from [rowId, rowId + count)
.
- getLoss() - Method in interface org.apache.spark.ml.param.shared.HasLoss
-
- getLoss() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getLossType() - Method in interface org.apache.spark.ml.tree.GBTClassifierParams
-
- getLossType() - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
-
- getLower() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
-
- getLowerBound(double, long, double) - Static method in class org.apache.spark.util.random.BinomialBounds
-
Returns a threshold p
such that if we conduct n Bernoulli trials with success rate = p
,
it is very unlikely to have more than fraction * n
successes.
- getLowerBound(double) - Static method in class org.apache.spark.util.random.PoissonBounds
-
Returns a lambda such that Pr[X > s] is very small, where X ~ Pois(lambda).
- getLowerBoundsOnCoefficients() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
- getLowerBoundsOnIntercepts() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
- getMap(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of map type as a Scala Map.
- getMap(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getMap(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the map type value for rowId
.
- getMapOutputMetadata() - Method in class org.apache.spark.shuffle.api.metadata.MapOutputCommitMessage
-
- getMapStatus(long) - Method in class org.apache.spark.ShuffleStatus
-
Get the map output that corresponding to a given mapId.
- GetMatchingBlockIds(Function1<BlockId, Object>, boolean) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
-
- GetMatchingBlockIds$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds$
-
- getMax() - Method in interface org.apache.spark.ml.feature.MinMaxScalerParams
-
- getMaxBins() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMaxBins() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxBlockSizeInMB() - Method in interface org.apache.spark.ml.param.shared.HasMaxBlockSizeInMB
-
- getMaxCategories() - Method in interface org.apache.spark.ml.feature.VectorIndexerParams
-
- getMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 max_cores = 4;
- getMaxCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional int32 max_cores = 4;
- getMaxCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional int32 max_cores = 4;
- getMaxDepth() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMaxDepth() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxDF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
-
- getMaxFailures(SparkConf, boolean) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
- getMaxIter() - Method in interface org.apache.spark.ml.param.shared.HasMaxIter
-
- getMaxIterations() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the max number of k-means iterations to split clusters.
- getMaxIterations() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Return the maximum number of iterations allowed
- getMaxIterations() - Method in class org.apache.spark.mllib.clustering.KMeans
-
Maximum number of iterations allowed.
- getMaxIterations() - Method in class org.apache.spark.mllib.clustering.LDA
-
Maximum number of iterations allowed.
- getMaxLocalProjDBSize() - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- getMaxLocalProjDBSize() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Gets the maximum number of items allowed in a projected database before local processing.
- getMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 max_memory = 19;
- getMaxMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 max_memory = 19;
- getMaxMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 max_memory = 19;
- getMaxMemoryInMB() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMaxMemoryInMB() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxPatternLength() - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- getMaxPatternLength() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Gets the maximal pattern length (i.e.
- getMaxSentenceLength() - Method in interface org.apache.spark.ml.feature.Word2VecBase
-
- getMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 max_tasks = 8;
- getMaxTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 max_tasks = 8;
- getMaxTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 max_tasks = 8;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double memory_bytes_spilled = 16;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double memory_bytes_spilled = 16;
- getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double memory_bytes_spilled = 16;
- getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 memory_bytes_spilled = 13;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 memory_bytes_spilled = 13;
- getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 memory_bytes_spilled = 13;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 memory_bytes_spilled = 21;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 memory_bytes_spilled = 21;
- getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 memory_bytes_spilled = 21;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 memory_bytes_spilled = 23;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 memory_bytes_spilled = 23;
- getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 memory_bytes_spilled = 23;
- getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilled(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilled(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 memory_bytes_spilled = 8;
- getMemoryBytesSpilled() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 memory_bytes_spilled = 8;
- getMemoryBytesSpilled() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 memory_bytes_spilled = 8;
- getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilledCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilledCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 14;
- getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilledList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double memory_bytes_spilled = 13;
- getMemoryBytesSpilledList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double memory_bytes_spilled = 13;
- getMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- getMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 memory_per_executor_mb = 6;
- getMemoryPerExecutorMb() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional int32 memory_per_executor_mb = 6;
- getMemoryPerExecutorMb() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional int32 memory_per_executor_mb = 6;
- getMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_remaining = 3;
- getMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
int64 memory_remaining = 3;
- getMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
int64 memory_remaining = 3;
- GetMemoryStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetMemoryStatus$
-
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 memory_used = 5;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 memory_used = 5;
- getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 memory_used = 5;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_used = 2;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
int64 memory_used = 2;
- getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
int64 memory_used = 2;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 memory_used = 3;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
int64 memory_used = 3;
- getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
int64 memory_used = 3;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 memory_used = 6;
- getMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
int64 memory_used = 6;
- getMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
int64 memory_used = 6;
- getMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 memory_used_bytes = 8;
- getMemoryUsedBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 memory_used_bytes = 8;
- getMemoryUsedBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 memory_used_bytes = 8;
- getMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 mem_size = 8;
- getMemSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
int64 mem_size = 8;
- getMemSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
int64 mem_size = 8;
- getMergedFetchFallbackCount(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCount(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCount(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double merged_fetch_fallback_count = 2;
- getMergedFetchFallbackCountList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double merged_fetch_fallback_count = 2;
- getMergerLocs() - Method in class org.apache.spark.ShuffleDependency
-
- getMessage(String, Map<String, String>) - Static method in class org.apache.spark.SparkThrowableHelper
-
- getMessage(String, Map<String, String>) - Static method in class org.apache.spark.SparkThrowableHelper
-
- getMessage(String, Map<String, String>, String) - Static method in class org.apache.spark.SparkThrowableHelper
-
- getMessage(SparkThrowable, Enumeration.Value) - Static method in class org.apache.spark.SparkThrowableHelper
-
- getMessage() - Method in exception org.apache.spark.sql.AnalysisException
-
- getMessageParameters() - Method in exception org.apache.spark.SparkException
-
- getMessageParameters() - Method in interface org.apache.spark.SparkThrowable
-
- getMessageParameters() - Method in exception org.apache.spark.sql.AnalysisException
-
- getMessageParameters() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
-
- getMessageTemplate(String) - Method in class org.apache.spark.ErrorClassesJsonReader
-
- getMetadata(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Metadata.
- getMetadataArray(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a Metadata array.
- getMetricDistribution(String) - Method in class org.apache.spark.status.api.v1.ExecutorPeakMetricsDistributions
-
Returns the distributions for the specified metric.
- getMetricLabel() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getMetricLabel() - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- getMetricName() - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
Get a BinaryClassificationMetrics, which can be used to get binary classification
metrics such as areaUnderROC and areaUnderPR.
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
Get a ClusteringMetrics, which can be used to get clustering metrics such as
silhouette score.
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
Get a MulticlassMetrics, which can be used to get multiclass classification
metrics such as accuracy, weightedPrecision, etc.
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
Get a MultilabelMetrics, which can be used to get multilabel classification
metrics such as accuracy, precision, precisionByLabel, etc.
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
Get a RankingMetrics, which can be used to get ranking metrics
such as meanAveragePrecision, meanAveragePrecisionAtK, etc.
- getMetrics(Dataset<?>) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
Get a RegressionMetrics, which can be used to get regression
metrics such as rootMeanSquaredError, meanSquaredError, etc.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
Deprecated.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
Deprecated.
- getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
Deprecated.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
Deprecated.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
Deprecated.
- getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
Deprecated.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
Deprecated.
- getMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
Deprecated.
- getMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
Deprecated.
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetrics(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetrics(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
map<string, int64> metrics = 1;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
map<string, string> metrics = 3;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
map<string, string> metrics = 8;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
map<string, int64> metrics = 1;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
map<string, int64> metrics = 1;
- getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
map<string, int64> metrics = 1;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
map<string, string> metrics = 3;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
map<string, string> metrics = 3;
- getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
map<string, string> metrics = 3;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
map<string, string> metrics = 8;
- getMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
map<string, string> metrics = 8;
- getMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
map<string, string> metrics = 8;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- getMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
map<string, int64> metrics = 1;
- getMetricsOrDefault(String, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
map<string, int64> metrics = 1;
- getMetricsOrDefault(String, long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
map<string, int64> metrics = 1;
- getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
map<string, string> metrics = 3;
- getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
map<string, string> metrics = 3;
- getMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
map<string, string> metrics = 3;
- getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
map<string, string> metrics = 8;
- getMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
map<string, string> metrics = 8;
- getMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
map<string, string> metrics = 8;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
map<string, int64> metrics = 1;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
map<string, int64> metrics = 1;
- getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder
-
map<string, int64> metrics = 1;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
map<string, string> metrics = 3;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
map<string, string> metrics = 3;
- getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
map<string, string> metrics = 3;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
map<string, string> metrics = 8;
- getMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
map<string, string> metrics = 8;
- getMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
map<string, string> metrics = 8;
- getMetricsProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- getMetricsSources(String) - Method in class org.apache.spark.BarrierTaskContext
-
- getMetricsSources(String) - Method in class org.apache.spark.TaskContext
-
::DeveloperApi::
Returns all metrics sources with the given name which are associated with the instance
which runs the task.
- getMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string metric_type = 3;
- getMetricType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
optional string metric_type = 3;
- getMetricType() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
-
optional string metric_type = 3;
- getMetricTypeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string metric_type = 3;
- getMetricTypeBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
optional string metric_type = 3;
- getMetricTypeBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
-
optional string metric_type = 3;
- getMetricValue(MemoryManager) - Method in interface org.apache.spark.metrics.SingleValueExecutorMetricType
-
- getMetricValues(MemoryManager) - Method in interface org.apache.spark.metrics.ExecutorMetricType
-
- getMetricValues(MemoryManager) - Method in interface org.apache.spark.metrics.SingleValueExecutorMetricType
-
- getMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
Deprecated.
- getMetricValues() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
Deprecated.
- getMetricValuesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getMetricValuesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getMetricValuesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, string> metric_values = 14;
- getMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
bool metric_values_is_null = 13;
- getMetricValuesIsNull() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
bool metric_values_is_null = 13;
- getMetricValuesIsNull() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
bool metric_values_is_null = 13;
- getMetricValuesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, string> metric_values = 14;
- getMetricValuesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, string> metric_values = 14;
- getMetricValuesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, string> metric_values = 14;
- getMetricValuesOrDefault(long, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, string> metric_values = 14;
- getMetricValuesOrDefault(long, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, string> metric_values = 14;
- getMetricValuesOrDefault(long, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, string> metric_values = 14;
- getMetricValuesOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<int64, string> metric_values = 14;
- getMetricValuesOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<int64, string> metric_values = 14;
- getMetricValuesOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<int64, string> metric_values = 14;
- getMin() - Method in interface org.apache.spark.ml.feature.MinMaxScalerParams
-
- getMinConfidence() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
-
- getMinCount() - Method in interface org.apache.spark.ml.feature.Word2VecBase
-
- getMinDF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
-
- getMinDivisibleClusterSize() - Method in interface org.apache.spark.ml.clustering.BisectingKMeansParams
-
- getMinDivisibleClusterSize() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the minimum number of points (if greater than or equal to 1.0
) or the minimum proportion
of points (if less than 1.0
) of a divisible cluster.
- getMinDocFreq() - Method in interface org.apache.spark.ml.feature.IDFBase
-
- getMiniBatchFraction() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
-
- getMiniBatchFraction() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Mini-batch fraction, which sets the fraction of document sampled and used in each iteration
- getMinInfoGain() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMinInfoGain() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMinInstancesPerNode() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMinInstancesPerNode() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMinSupport() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
-
- getMinSupport() - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- getMinSupport() - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Get the minimal support (i.e.
- getMinTF() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
-
- getMinTokenLength() - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- getMinWeightFractionPerNode() - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- getMinWeightFractionPerNode() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getMissingValue() - Method in interface org.apache.spark.ml.feature.ImputerParams
-
- getMode(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the OpenCV representation as an int
- getModelType() - Method in interface org.apache.spark.ml.classification.NaiveBayesParams
-
- getModelType() - Method in class org.apache.spark.mllib.classification.NaiveBayes
-
Get the model type.
- getModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
Deprecated.
- getModifiedConfigs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
Deprecated.
- getModifiedConfigsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- getModifiedConfigsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getModifiedConfigsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<string, string> modified_configs = 6;
- getModifiedConfigsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
map<string, string> modified_configs = 6;
- getModifiedConfigsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
map<string, string> modified_configs = 6;
- getMutableAttributes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getMutableCustomMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
Deprecated.
- getMutableDurationMs() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getMutableEventTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getMutableExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getMutableExecutorLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
Deprecated.
- getMutableExecutorResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
Deprecated.
- getMutableExecutorSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getMutableJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getMutableJobsValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getMutableKilledTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getMutableKillTasksSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
Deprecated.
- getMutableLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
Deprecated.
- getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
Deprecated.
- getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
Deprecated.
- getMutableMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
Deprecated.
- getMutableMetricValues() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getMutableModifiedConfigs() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
Deprecated.
- getMutableObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getMutableProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
Deprecated.
- getMutableResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getMutableTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
Deprecated.
- getMutableTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getN() - Method in class org.apache.spark.ml.feature.NGram
-
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
optional string name = 1;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
optional string name = 1;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
optional string name = 2;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
optional string name = 2;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
optional string name = 1;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string name = 39;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string name = 39;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string name = 39;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string name = 1;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string name = 1;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string name = 1;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string name = 3;
- getName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string name = 3;
- getName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string name = 3;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
optional string name = 1;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapperOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
optional string name = 1;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceInformationOrBuilder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
optional string name = 2;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeOrBuilder
-
optional string name = 2;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
optional string name = 1;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetricOrBuilder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string name = 39;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string name = 39;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string name = 39;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string name = 1;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string name = 1;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string name = 1;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string name = 3;
- getNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string name = 3;
- getNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string name = 3;
- getNames() - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- getNChannels(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the number of channels in the image
- getNode(int, Node) - Static method in class org.apache.spark.mllib.tree.model.Node
-
Traces down from a root node to get the node with the given node index.
- getNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNode() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNode() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNodeBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNodeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNodeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNodeOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNodesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- getNonnegative() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_stages = 16;
- getNumActiveStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_active_stages = 16;
- getNumActiveStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_active_stages = 16;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_tasks = 10;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_active_tasks = 10;
- getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_active_tasks = 10;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_active_tasks = 2;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
int32 num_active_tasks = 2;
- getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
-
int32 num_active_tasks = 2;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_active_tasks = 5;
- getNumActiveTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_active_tasks = 5;
- getNumActiveTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_active_tasks = 5;
- getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
-
- getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
-
- getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase
-
- getNumber() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
- getNumBins() - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getNumBins() - Method in class org.apache.spark.sql.util.NumericHistogram
-
Returns the number of bins.
- getNumBuckets() - Method in interface org.apache.spark.ml.feature.QuantileDiscretizerBase
-
- getNumBucketsArray() - Method in interface org.apache.spark.ml.feature.QuantileDiscretizerBase
-
- getNumBytesWritten() - Method in interface org.apache.spark.shuffle.api.ShufflePartitionWriter
-
- getNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_cached_partitions = 4;
- getNumCachedPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
int32 num_cached_partitions = 4;
- getNumCachedPartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
int32 num_cached_partitions = 4;
- getNumClasses(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
-
Examine a schema to identify the number of classes in a label column.
- getNumClasses() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_indices = 15;
- getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_completed_indices = 15;
- getNumCompletedIndices() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_completed_indices = 15;
- getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_completed_indices = 9;
- getNumCompletedIndices() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_completed_indices = 9;
- getNumCompletedIndices() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_completed_indices = 9;
- getNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_jobs = 1;
- getNumCompletedJobs() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
int32 num_completed_jobs = 1;
- getNumCompletedJobs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AppSummaryOrBuilder
-
int32 num_completed_jobs = 1;
- getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_stages = 2;
- getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
int32 num_completed_stages = 2;
- getNumCompletedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AppSummaryOrBuilder
-
int32 num_completed_stages = 2;
- getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_stages = 17;
- getNumCompletedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_completed_stages = 17;
- getNumCompletedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_completed_stages = 17;
- getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_tasks = 11;
- getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_completed_tasks = 11;
- getNumCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_completed_tasks = 11;
- getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_completed_tasks = 3;
- getNumCompletedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
int32 num_completed_tasks = 3;
- getNumCompletedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
-
int32 num_completed_tasks = 3;
- getNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_complete_tasks = 6;
- getNumCompleteTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_complete_tasks = 6;
- getNumCompleteTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_complete_tasks = 6;
- getNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_stages = 19;
- getNumFailedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_failed_stages = 19;
- getNumFailedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_failed_stages = 19;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_tasks = 13;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_failed_tasks = 13;
- getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_failed_tasks = 13;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_failed_tasks = 4;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
int32 num_failed_tasks = 4;
- getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
-
int32 num_failed_tasks = 4;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_failed_tasks = 7;
- getNumFailedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_failed_tasks = 7;
- getNumFailedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_failed_tasks = 7;
- getNumFeatures() - Method in interface org.apache.spark.ml.param.shared.HasNumFeatures
-
- getNumFeatures(StructField) - Static method in class org.apache.spark.ml.util.MetadataUtils
-
Examine a schema to identify the number of features in a vector column.
- getNumFeatures() - Method in class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
The dimension of training features.
- getNumFolds() - Method in interface org.apache.spark.ml.tuning.CrossValidatorParams
-
- getNumHashTables() - Method in interface org.apache.spark.ml.feature.LSHParams
-
- getNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
int64 num_input_rows = 5;
- getNumInputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
int64 num_input_rows = 5;
- getNumInputRows() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
int64 num_input_rows = 5;
- getNumItemBlocks() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getNumIterations() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_killed_tasks = 14;
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_killed_tasks = 14;
- getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_killed_tasks = 14;
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_killed_tasks = 5;
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
int32 num_killed_tasks = 5;
- getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
-
int32 num_killed_tasks = 5;
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_killed_tasks = 8;
- getNumKilledTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_killed_tasks = 8;
- getNumKilledTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_killed_tasks = 8;
- getNumObjFields() - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
int64 num_output_rows = 2;
- getNumOutputRows() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
int64 num_output_rows = 2;
- getNumOutputRows() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder
-
int64 num_output_rows = 2;
- getNumPartitions() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return the number of partitions in this RDD.
- getNumPartitions() - Method in interface org.apache.spark.ml.feature.Word2VecBase
-
- getNumPartitions() - Method in interface org.apache.spark.ml.fpm.FPGrowthParams
-
- getNumPartitions() - Method in class org.apache.spark.rdd.RDD
-
Returns the number of partitions of this RDD.
- getNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_partitions = 3;
- getNumPartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
int32 num_partitions = 3;
- getNumPartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
int32 num_partitions = 3;
- getNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_dropped_by_watermark = 9;
- getNumRowsDroppedByWatermark() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_rows_dropped_by_watermark = 9;
- getNumRowsDroppedByWatermark() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_rows_dropped_by_watermark = 9;
- getNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_removed = 5;
- getNumRowsRemoved() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_rows_removed = 5;
- getNumRowsRemoved() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_rows_removed = 5;
- getNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_total = 2;
- getNumRowsTotal() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_rows_total = 2;
- getNumRowsTotal() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_rows_total = 2;
- getNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_updated = 3;
- getNumRowsUpdated() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_rows_updated = 3;
- getNumRowsUpdated() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_rows_updated = 3;
- getNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_shuffle_partitions = 10;
- getNumShufflePartitions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_shuffle_partitions = 10;
- getNumShufflePartitions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_shuffle_partitions = 10;
- getNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_stages = 18;
- getNumSkippedStages() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_skipped_stages = 18;
- getNumSkippedStages() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_skipped_stages = 18;
- getNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_tasks = 12;
- getNumSkippedTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_skipped_tasks = 12;
- getNumSkippedTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_skipped_tasks = 12;
- getNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_state_store_instances = 11;
- getNumStateStoreInstances() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
int64 num_state_store_instances = 11;
- getNumStateStoreInstances() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
int64 num_state_store_instances = 11;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_tasks = 9;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
int32 num_tasks = 9;
- getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
int32 num_tasks = 9;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_tasks = 1;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
int32 num_tasks = 1;
- getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryOrBuilder
-
int32 num_tasks = 1;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_tasks = 4;
- getNumTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 num_tasks = 4;
- getNumTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 num_tasks = 4;
- getNumTopFeatures() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getNumTrees() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
Number of trees in ensemble
- getNumTrees() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
Number of trees in ensemble
- getNumTrees() - Method in interface org.apache.spark.ml.tree.RandomForestParams
-
- getNumUserBlocks() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getNumValues() - Method in class org.apache.spark.ml.attribute.NominalAttribute
-
Get the number of values, either from numValues
or from values
.
- getObjFieldValues(Object, Object[]) - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
Deprecated.
- getObservedMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
Deprecated.
- getObservedMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
Deprecated.
- getObservedMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- getObservedMetricsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getObservedMetricsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> observed_metrics = 12;
- getObservedMetricsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
map<string, string> observed_metrics = 12;
- getObservedMetricsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
map<string, string> observed_metrics = 12;
- getOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_remaining = 8;
- getOffHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional int64 off_heap_memory_remaining = 8;
- getOffHeapMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional int64 off_heap_memory_remaining = 8;
- getOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_used = 6;
- getOffHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional int64 off_heap_memory_used = 6;
- getOffHeapMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional int64 off_heap_memory_used = 6;
- getOffset() - Method in interface org.apache.spark.sql.connector.read.streaming.ContinuousPartitionReader
-
Get the offset of the current record, or the start offset if no records have been read.
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getOffsetClause(Integer) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns the OFFSET clause for the SELECT statement
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getOffsetClause(Integer) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getOffsetCol() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getOldBoostingStrategy(Map<Object, Object>, Enumeration.Value) - Method in interface org.apache.spark.ml.tree.GBTParams
-
(private[ml]) Create a BoostingStrategy instance to use with the old API.
- getOldDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
Get docConcentration used by spark.mllib LDA
- getOldImpurity() - Method in interface org.apache.spark.ml.tree.HasVarianceImpurity
-
Convert new impurity to old impurity.
- getOldImpurity() - Method in interface org.apache.spark.ml.tree.TreeClassifierParams
-
Convert new impurity to old impurity.
- getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTClassifierParams
-
(private[ml]) Convert new loss to old loss.
- getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTParams
-
Get old Gradient Boosting Loss type
- getOldLossType() - Method in interface org.apache.spark.ml.tree.GBTRegressorParams
-
(private[ml]) Convert new loss to old loss.
- getOldOptimizer() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getOldStrategy(Map<Object, Object>, int, Enumeration.Value, Impurity, double) - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
(private[ml]) Create a Strategy instance to use with the old API.
- getOldStrategy(Map<Object, Object>, int, Enumeration.Value, Impurity) - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
-
Create a Strategy instance to use with the old API.
- getOldTopicConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
Get topicConcentration used by spark.mllib LDA
- getOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_remaining = 7;
- getOnHeapMemoryRemaining() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional int64 on_heap_memory_remaining = 7;
- getOnHeapMemoryRemaining() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional int64 on_heap_memory_remaining = 7;
- getOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_used = 5;
- getOnHeapMemoryUsed() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
optional int64 on_heap_memory_used = 5;
- getOnHeapMemoryUsed() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDDataDistributionOrBuilder
-
optional int64 on_heap_memory_used = 5;
- getOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
optional string operator_name = 1;
- getOperatorName() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
optional string operator_name = 1;
- getOperatorName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
optional string operator_name = 1;
- getOperatorNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
optional string operator_name = 1;
- getOperatorNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
optional string operator_name = 1;
- getOperatorNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder
-
optional string operator_name = 1;
- getOptimizeDocConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getOptimizeDocConcentration() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Optimize docConcentration, indicates whether docConcentration (Dirichlet parameter for
document-topic distribution) will be optimized during training.
- getOptimizer() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getOptimizer() - Method in class org.apache.spark.mllib.clustering.LDA
-
LDAOptimizer used to perform the actual calculation
- getOption(String) - Method in class org.apache.spark.SparkConf
-
Get a parameter as an Option
- getOption(String) - Method in class org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- getOption() - Method in interface org.apache.spark.sql.streaming.GroupState
-
Get the state value as a scala Option.
- getOption() - Method in class org.apache.spark.streaming.State
-
Get the state as a scala.Option
.
- getOptional(boolean, Function0<T>) - Static method in class org.apache.spark.status.protobuf.Utils
-
- getOrCreate(SparkConf) - Static method in class org.apache.spark.SparkContext
-
This function may be used to get or instantiate a SparkContext and register it as a
singleton object.
- getOrCreate() - Static method in class org.apache.spark.SparkContext
-
This function may be used to get or instantiate a SparkContext and register it as a
singleton object.
- getOrCreate() - Method in class org.apache.spark.sql.SparkSession.Builder
-
Gets an existing
SparkSession
or, if there is no existing one, creates a new
one based on the options set in this builder.
- getOrCreate(SparkContext) - Static method in class org.apache.spark.sql.SQLContext
-
- getOrCreate(String, Function0<JavaStreamingContext>) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<JavaStreamingContext>, Configuration) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<JavaStreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - Static method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreateSparkSession(JavaSparkContext, Map<Object, Object>, boolean) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- getOrDefault(Param<T>) - Method in interface org.apache.spark.ml.param.Params
-
Gets the value of a param in the embedded param map or its default value.
- getOrDiscoverAllResources(SparkConf, String, Option<String>) - Static method in class org.apache.spark.resource.ResourceUtils
-
Gets all allocated resource information for the input component from input resources file and
the application level Spark configs.
- getOrDiscoverAllResourcesForResourceProfile(Option<String>, String, ResourceProfile, SparkConf) - Static method in class org.apache.spark.resource.ResourceUtils
-
This function is similar to getOrDiscoverallResources, except for it uses the ResourceProfile
information instead of the application level configs.
- getOrElse(Param<T>, T) - Method in class org.apache.spark.ml.param.ParamMap
-
Returns the value associated with a param or a default value.
- getOrigin(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the origin of the image
- getOutgoingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdges(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdges(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutgoingEdgesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- getOutputAttrGroupFromData(Dataset<?>, Seq<String>, Seq<String>, boolean) - Static method in class org.apache.spark.ml.feature.OneHotEncoderCommon
-
This method is called when we want to generate AttributeGroup
from actual data for
one-hot encoder.
- getOutputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- getOutputBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_bytes = 8;
- getOutputBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_bytes = 8;
- getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_bytes = 7;
- getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 output_bytes = 7;
- getOutputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 output_bytes = 7;
- getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_bytes = 26;
- getOutputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 output_bytes = 26;
- getOutputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 output_bytes = 26;
- getOutputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- getOutputBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_bytes = 8;
- getOutputBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_bytes = 8;
- getOutputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- getOutputBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_bytes = 8;
- getOutputBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_bytes = 8;
- getOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_bytes_written = 28;
- getOutputBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 output_bytes_written = 28;
- getOutputBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 output_bytes_written = 28;
- getOutputCol() - Method in interface org.apache.spark.ml.param.shared.HasOutputCol
-
- getOutputCols() - Method in interface org.apache.spark.ml.param.shared.HasOutputCols
-
- getOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputDeterministicLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputDeterministicLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputDeterministicLevelValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputDeterministicLevelValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputDeterministicLevelValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationNodeOrBuilder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- getOutputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- getOutputRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_records = 9;
- getOutputRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_records = 9;
- getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_records = 8;
- getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 output_records = 8;
- getOutputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 output_records = 8;
- getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_records = 27;
- getOutputRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 output_records = 27;
- getOutputRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 output_records = 27;
- getOutputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- getOutputRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_records = 9;
- getOutputRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_records = 9;
- getOutputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- getOutputRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double output_records = 9;
- getOutputRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double output_records = 9;
- getOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_records_written = 29;
- getOutputRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 output_records_written = 29;
- getOutputRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 output_records_written = 29;
- getOutputSize(int) - Method in interface org.apache.spark.ml.ann.Layer
-
Returns the output size given the input size (not counting the stack size).
- getOutputStream(String, Configuration) - Static method in class org.apache.spark.streaming.util.HdfsUtils
-
- getP() - Method in class org.apache.spark.ml.feature.Normalizer
-
- getParallelism() - Method in interface org.apache.spark.ml.param.shared.HasParallelism
-
- getParam(String) - Method in interface org.apache.spark.ml.param.Params
-
Gets a param by its name.
- getParameter(String) - Method in class org.apache.spark.ui.XssSafeRequest
-
- getParameterMap() - Method in class org.apache.spark.ui.XssSafeRequest
-
- getParameterNames() - Method in class org.apache.spark.ui.XssSafeRequest
-
- getParameterOtherTable(HttpServletRequest, String) - Method in interface org.apache.spark.ui.PagedTable
-
Returns parameters of other tables in the page.
- getParameterValues(String) - Method in class org.apache.spark.ui.XssSafeRequest
-
- getParentLoggerNotImplementedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- getParents(int) - Method in class org.apache.spark.NarrowDependency
-
Get the parent partitions for a child partition.
- getParents(int) - Method in class org.apache.spark.OneToOneDependency
-
- getParents(int) - Method in class org.apache.spark.RangeDependency
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- getParserForType() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
-
- getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.EdgePartition1D$
-
- getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.EdgePartition2D$
-
- getPartition(long, long, int) - Method in interface org.apache.spark.graphx.PartitionStrategy
-
Returns the partition number for a given edge.
- getPartition(long, long, int) - Method in class org.apache.spark.graphx.PartitionStrategy.RandomVertexCut$
-
- getPartition(Object) - Method in class org.apache.spark.HashPartitioner
-
- getPartition(Object) - Method in class org.apache.spark.Partitioner
-
- getPartition(Object) - Method in class org.apache.spark.RangePartitioner
-
- getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 partition_id = 4;
- getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int32 partition_id = 4;
- getPartitionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int32 partition_id = 4;
- getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 partition_id = 4;
- getPartitionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int32 partition_id = 4;
- getPartitionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int32 partition_id = 4;
- getPartitionId() - Static method in class org.apache.spark.TaskContext
-
Returns the partition id of currently active TaskContext.
- getPartitionLengths() - Method in class org.apache.spark.shuffle.api.metadata.MapOutputCommitMessage
-
- getPartitionMetadataByFilterError(InvocationTargetException) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- getPartitions() - Method in class org.apache.spark.api.r.BaseRRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.CoGroupedRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
- getPartitions() - Method in class org.apache.spark.rdd.HadoopRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.JdbcRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.NewHadoopRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.ShuffledRDD
-
- getPartitions() - Method in class org.apache.spark.rdd.UnionRDD
-
- getPartitions() - Method in class org.apache.spark.status.LiveRDD
-
- getPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitions(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- getPartitionWriter(int) - Method in interface org.apache.spark.shuffle.api.ShuffleMapOutputWriter
-
Creates a writer that can open an output stream to persist bytes targeted for a given reduce
partition id.
- getPath() - Method in class org.apache.spark.input.PortableDataStream
-
- getPattern() - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double peak_execution_memory = 15;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double peak_execution_memory = 15;
- getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double peak_execution_memory = 15;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 peak_execution_memory = 23;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 peak_execution_memory = 23;
- getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 peak_execution_memory = 23;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 peak_execution_memory = 25;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 peak_execution_memory = 25;
- getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 peak_execution_memory = 25;
- getPeakExecutionMemory(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemory(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemory(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 peak_execution_memory = 10;
- getPeakExecutionMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 peak_execution_memory = 10;
- getPeakExecutionMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 peak_execution_memory = 10;
- getPeakExecutionMemoryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemoryCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemoryCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemoryList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemoryList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double peak_execution_memory = 12;
- getPeakExecutionMemoryList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double peak_execution_memory = 12;
- getPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakExecutorMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- getPeakMemoryMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- GetPeers(BlockManagerId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetPeers
-
- GetPeers$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetPeers$
-
- getPercentile() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getPersistentRDDs() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
- getPersistentRDDs() - Method in class org.apache.spark.SparkContext
-
Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.
- getPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string physical_plan_description = 5;
- getPhysicalPlanDescription() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string physical_plan_description = 5;
- getPhysicalPlanDescription() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string physical_plan_description = 5;
- getPhysicalPlanDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string physical_plan_description = 5;
- getPhysicalPlanDescriptionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
optional string physical_plan_description = 5;
- getPhysicalPlanDescriptionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
optional string physical_plan_description = 5;
- getPmml() - Method in interface org.apache.spark.mllib.pmml.export.PMMLModelExport
-
- getPoissonSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long, ClassTag<K>, ClassTag<V>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Return the per partition sampling function used for sampling with replacement.
- getPoolForName(String) - Method in class org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return the pool associated with the given name, if one exists
- getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.Latest
-
- getPosition() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.TrimHorizon
-
- getPowerIterationClustering(int, String, int, String, String, String) - Static method in class org.apache.spark.ml.r.PowerIterationClusteringWrapper
-
- getPredictionCol() - Method in interface org.apache.spark.ml.param.shared.HasPredictionCol
-
- getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.HadoopRDD
-
- getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.NewHadoopRDD
-
- getPreferredLocations(Partition) - Method in class org.apache.spark.rdd.UnionRDD
-
- getPrefixSpan(double, int, double, String) - Static method in class org.apache.spark.ml.r.PrefixSpanWrapper
-
- getProbabilityCol() - Method in interface org.apache.spark.ml.param.shared.HasProbabilityCol
-
- getProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double processed_rows_per_second = 7;
- getProcessedRowsPerSecond() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
double processed_rows_per_second = 7;
- getProcessedRowsPerSecond() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
double processed_rows_per_second = 7;
- getProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
Deprecated.
- getProcessLogs() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
Deprecated.
- getProcessLogs() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
Deprecated.
- getProcessLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- getProcessLogsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getProcessLogsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
map<string, string> process_logs = 7;
- getProcessLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
map<string, string> process_logs = 7;
- getProcessLogsMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
map<string, string> process_logs = 7;
- getProcessLogsMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
map<string, string> process_logs = 7;
- getProcessLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
map<string, string> process_logs = 7;
- getProcessLogsOrDefault(String, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
map<string, string> process_logs = 7;
- getProcessLogsOrDefault(String, String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
map<string, string> process_logs = 7;
- getProcessLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
map<string, string> process_logs = 7;
- getProcessLogsOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
map<string, string> process_logs = 7;
- getProcessLogsOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
map<string, string> process_logs = 7;
- getProcessName() - Static method in class org.apache.spark.util.Utils
-
Returns the name of this JVM process.
- getProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgress() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgress() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapperOrBuilder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgressBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgressOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgressOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getProgressOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapperOrBuilder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- getPropertiesFromFile(String) - Static method in class org.apache.spark.util.Utils
-
Load properties present in the given file.
- getQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
optional string quantile = 3;
- getQuantile() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
optional string quantile = 3;
- getQuantile() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
optional string quantile = 3;
- getQuantileBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
optional string quantile = 3;
- getQuantileBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
optional string quantile = 3;
- getQuantileBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
optional string quantile = 3;
- getQuantileCalculationStrategy() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getQuantileProbabilities() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
-
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double quantiles = 1;
- getQuantiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesCol() - Method in interface org.apache.spark.ml.regression.AFTSurvivalRegressionParams
-
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double quantiles = 1;
- getQuantilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double quantiles = 1;
- getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
repeated double quantiles = 1;
- getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- getQuantilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double quantiles = 1;
- getQuantilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double quantiles = 1;
- getQuantilesValue(IndexedSeq<Object>, double[]) - Static method in class org.apache.spark.status.AppStatusUtils
-
- getQueryContext() - Method in exception org.apache.spark.SparkException
-
- getQueryContext() - Method in interface org.apache.spark.SparkThrowable
-
- getQueryContext() - Method in exception org.apache.spark.sql.AnalysisException
-
- getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- getQueryContext(SQLQueryContext) - Method in interface org.apache.spark.sql.errors.DataTypeErrorsBase
-
- getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- getQueryContext(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- getQueryName(StreamingQueryUIData) - Static method in class org.apache.spark.sql.streaming.ui.UIUtils
-
- getQueryStatus(StreamingQueryUIData) - Static method in class org.apache.spark.sql.streaming.ui.UIUtils
-
- getRandomSample(Seq<T>, int, Random) - Static method in class org.apache.spark.storage.BlockReplicationUtils
-
Get a random sample of size m from the elems
- getRank() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getRatingCol() - Method in interface org.apache.spark.ml.recommendation.ALSParams
-
- getRawPredictionCol() - Method in interface org.apache.spark.ml.param.shared.HasRawPredictionCol
-
- getRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 rdd_blocks = 4;
- getRddBlocks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 rdd_blocks = 4;
- getRddBlocks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 rdd_blocks = 4;
- GetRDDBlockVisibility(RDDBlockId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetRDDBlockVisibility
-
- GetRDDBlockVisibility$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetRDDBlockVisibility$
-
- getRddIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- getRddIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated int64 rdd_ids = 43;
- getRddIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated int64 rdd_ids = 43;
- getRddIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- getRddIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated int64 rdd_ids = 43;
- getRddIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated int64 rdd_ids = 43;
- getRddIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- getRddIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
repeated int64 rdd_ids = 43;
- getRddIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
repeated int64 rdd_ids = 43;
- getRDDStorageInfo() - Method in class org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return information about what RDDs are cached, if they are in mem or on disk, how much space
they take, etc.
- getReadBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- getReadBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_bytes = 1;
- getReadBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_bytes = 1;
- getReadBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- getReadBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_bytes = 1;
- getReadBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_bytes = 1;
- getReadBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- getReadBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_bytes = 1;
- getReadBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_bytes = 1;
- getReadLimits() - Method in class org.apache.spark.sql.connector.read.streaming.CompositeReadLimit
-
- getReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- getReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_records = 2;
- getReadRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_records = 2;
- getReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- getReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_records = 2;
- getReadRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_records = 2;
- getReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- getReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double read_records = 2;
- getReadRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double read_records = 2;
- getReceiver() - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
-
Gets the receiver object that will be sent to the worker nodes
to receive data.
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_read = 19;
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double records_read = 19;
- getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double records_read = 19;
- getRecordsRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- getRecordsRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double records_read = 2;
- getRecordsRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double records_read = 2;
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 records_read = 2;
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
int64 records_read = 2;
- getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricsOrBuilder
-
int64 records_read = 2;
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 records_read = 7;
- getRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 records_read = 7;
- getRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 records_read = 7;
- getRecordsReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- getRecordsReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double records_read = 2;
- getRecordsReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double records_read = 2;
- getRecordsReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- getRecordsReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
repeated double records_read = 2;
- getRecordsReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributionsOrBuilder
-
repeated double records_read = 2;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_written = 21;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double records_written = 21;
- getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double records_written = 21;
- getRecordsWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- getRecordsWritten(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double records_written = 2;
- getRecordsWritten(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double records_written = 2;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 records_written = 2;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
int64 records_written = 2;
- getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricsOrBuilder
-
int64 records_written = 2;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 records_written = 3;
- getRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
int64 records_written = 3;
- getRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
-
int64 records_written = 3;
- getRecordsWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- getRecordsWrittenCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double records_written = 2;
- getRecordsWrittenCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double records_written = 2;
- getRecordsWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- getRecordsWrittenList() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
repeated double records_written = 2;
- getRecordsWrittenList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributionsOrBuilder
-
repeated double records_written = 2;
- getRegParam() - Method in interface org.apache.spark.ml.param.shared.HasRegParam
-
- getRelativeError() - Method in interface org.apache.spark.ml.param.shared.HasRelativeError
-
- getRemoteBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_blocks_fetched = 1;
- getRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 remote_blocks_fetched = 1;
- getRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 remote_blocks_fetched = 1;
- getRemoteBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_blocks_fetched = 3;
- getRemoteBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_blocks_fetched = 3;
- getRemoteBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read = 6;
- getRemoteBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read = 4;
- getRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 remote_bytes_read = 4;
- getRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 remote_bytes_read = 4;
- getRemoteBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read = 6;
- getRemoteBytesReadToDisk(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDisk(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDisk(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read_to_disk = 5;
- getRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 remote_bytes_read_to_disk = 5;
- getRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 remote_bytes_read_to_disk = 5;
- getRemoteBytesReadToDiskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDiskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDiskCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDiskList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDiskList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteBytesReadToDiskList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_bytes_read_to_disk = 7;
- getRemoteMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_blocks_fetched = 3;
- getRemoteMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_bytes_read = 7;
- getRemoteMergedBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 remote_merged_bytes_read = 7;
- getRemoteMergedBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedBytesReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_bytes_read = 7;
- getRemoteMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedChunksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_chunks_fetched = 5;
- getRemoteMergedReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
int64 remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricsOrBuilder
-
int64 remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteMergedReqsDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributionsOrBuilder
-
repeated double remote_merged_reqs_duration = 9;
- getRemoteReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDuration(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDuration(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_reqs_duration = 8;
- getRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
int64 remote_reqs_duration = 8;
- getRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
int64 remote_reqs_duration = 8;
- getRemoteReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDurationCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDurationCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDurationList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double remote_reqs_duration = 9;
- getRemoteReqsDurationList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double remote_reqs_duration = 9;
- getRemoteUser() - Method in class org.apache.spark.ui.XssSafeRequest
-
- getRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string remove_reason = 22;
- getRemoveReason() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string remove_reason = 22;
- getRemoveReason() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string remove_reason = 22;
- getRemoveReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string remove_reason = 22;
- getRemoveReasonBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional string remove_reason = 22;
- getRemoveReasonBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional string remove_reason = 22;
- getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional int64 remove_time = 21;
- getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
optional int64 remove_time = 21;
- getRemoveTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
optional int64 remove_time = 21;
- getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional int64 remove_time = 6;
- getRemoveTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
optional int64 remove_time = 6;
- getRemoveTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
optional int64 remove_time = 6;
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getRenameColumnQuery(String, String, String, int) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getRenameColumnQuery(String, String, String, int) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- GetReplicateInfoForRDDBlocks(BlockManagerId) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetReplicateInfoForRDDBlocks
-
- GetReplicateInfoForRDDBlocks$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetReplicateInfoForRDDBlocks$
-
- getResource(String) - Method in class org.apache.spark.util.ChildFirstURLClassLoader
-
- getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string resource_name = 1;
- getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string resource_name = 1;
- getResourceName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string resource_name = 1;
- getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
optional string resource_name = 1;
- getResourceName() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
optional string resource_name = 1;
- getResourceName() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
optional string resource_name = 1;
- getResourceNameBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequestOrBuilder
-
optional string resource_name = 1;
- getResourceProfile() - Method in class org.apache.spark.api.java.JavaRDD
-
Get the ResourceProfile specified with this RDD or None if it wasn't specified.
- getResourceProfile() - Method in class org.apache.spark.rdd.RDD
-
Get the ResourceProfile specified with this RDD or null if it wasn't specified.
- getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 resource_profile_id = 29;
- getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 resource_profile_id = 29;
- getResourceProfileId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 resource_profile_id = 29;
- getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 resource_profile_id = 49;
- getResourceProfileId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 resource_profile_id = 49;
- getResourceProfileId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 resource_profile_id = 49;
- getResourceProfiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfiles(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfiles(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResourceProfilesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- getResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
Deprecated.
- getResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
Deprecated.
- getResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
Deprecated.
- getResources(String) - Method in class org.apache.spark.util.ChildFirstURLClassLoader
-
- getResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- getResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrDefault(String, StoreTypes.ResourceInformation) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
map<string, .org.apache.spark.status.protobuf.ResourceInformation> resources = 28;
- getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 result_fetch_start = 6;
- getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional int64 result_fetch_start = 6;
- getResultFetchStart() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional int64 result_fetch_start = 6;
- getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_fetch_start = 6;
- getResultFetchStart() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 result_fetch_start = 6;
- getResultFetchStart() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 result_fetch_start = 6;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_serialization_time = 12;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double result_serialization_time = 12;
- getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double result_serialization_time = 12;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_serialization_time = 20;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 result_serialization_time = 20;
- getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 result_serialization_time = 20;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_serialization_time = 22;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 result_serialization_time = 22;
- getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 result_serialization_time = 22;
- getResultSerializationTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- getResultSerializationTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_serialization_time = 9;
- getResultSerializationTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_serialization_time = 9;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_serialization_time = 7;
- getResultSerializationTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 result_serialization_time = 7;
- getResultSerializationTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 result_serialization_time = 7;
- getResultSerializationTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- getResultSerializationTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_serialization_time = 9;
- getResultSerializationTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_serialization_time = 9;
- getResultSerializationTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- getResultSerializationTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_serialization_time = 9;
- getResultSerializationTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_serialization_time = 9;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_size = 10;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double result_size = 10;
- getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double result_size = 10;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_size = 18;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 result_size = 18;
- getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 result_size = 18;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_size = 20;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 result_size = 20;
- getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 result_size = 20;
- getResultSize(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- getResultSize(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_size = 7;
- getResultSize(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_size = 7;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_size = 5;
- getResultSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
int64 result_size = 5;
- getResultSize() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
int64 result_size = 5;
- getResultSizeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- getResultSizeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_size = 7;
- getResultSizeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_size = 7;
- getResultSizeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- getResultSizeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double result_size = 7;
- getResultSizeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double result_size = 7;
- getRoaringBitMap(ShuffleBlockChunkId) - Method in class org.apache.spark.storage.PushBasedFetchHelper
-
Get the RoaringBitMap for a specific ShuffleBlockChunkId
- getRollingIntervalSecs(SparkConf, boolean) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
- getRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootCluster() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootCluster() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootClusterBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootClusterOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootClusterOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- getRootDirectory() - Static method in class org.apache.spark.SparkFiles
-
Get the root directory that contains files added through SparkContext.addFile()
.
- getRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 root_execution_id = 2;
- getRootExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
int64 root_execution_id = 2;
- getRootExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
int64 root_execution_id = 2;
- getRow(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
-
Returns the row in this batch at `rowId`.
- getRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfo() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfo() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfoBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfoOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRpInfoOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapperOrBuilder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string run_id = 3;
- getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string run_id = 3;
- getRunId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string run_id = 3;
- getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string run_id = 2;
- getRunId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string run_id = 2;
- getRunId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string run_id = 2;
- getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string run_id = 3;
- getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
optional string run_id = 3;
- getRunIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
optional string run_id = 3;
- getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string run_id = 2;
- getRunIdBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string run_id = 2;
- getRunIdBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string run_id = 2;
- getRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntimeBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntimeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntimeOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getRuntimeOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- getScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string scala_version = 3;
- getScalaVersion() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string scala_version = 3;
- getScalaVersion() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string scala_version = 3;
- getScalaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string scala_version = 3;
- getScalaVersionBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
optional string scala_version = 3;
- getScalaVersionBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RuntimeInfoOrBuilder
-
optional string scala_version = 3;
- getScalingVec() - Method in class org.apache.spark.ml.feature.ElementwiseProduct
-
- getSchedulableByName(String) - Method in interface org.apache.spark.scheduler.Schedulable
-
- getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double scheduler_delay = 14;
- getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double scheduler_delay = 14;
- getSchedulerDelay() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double scheduler_delay = 14;
- getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 scheduler_delay = 17;
- getSchedulerDelay() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int64 scheduler_delay = 17;
- getSchedulerDelay() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int64 scheduler_delay = 17;
- getSchedulerDelay(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- getSchedulerDelay(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double scheduler_delay = 11;
- getSchedulerDelay(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double scheduler_delay = 11;
- getSchedulerDelayCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- getSchedulerDelayCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double scheduler_delay = 11;
- getSchedulerDelayCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double scheduler_delay = 11;
- getSchedulerDelayList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- getSchedulerDelayList() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
repeated double scheduler_delay = 11;
- getSchedulerDelayList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
repeated double scheduler_delay = 11;
- getSchedulingMode() - Method in class org.apache.spark.SparkContext
-
Return current scheduling mode
- getSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string scheduling_pool = 42;
- getSchedulingPool() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string scheduling_pool = 42;
- getSchedulingPool() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string scheduling_pool = 42;
- getSchedulingPoolBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string scheduling_pool = 42;
- getSchedulingPoolBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional string scheduling_pool = 42;
- getSchedulingPoolBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional string scheduling_pool = 42;
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getSchemaCommentQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getSchemaCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getSchemaQuery(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getSchemaQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
The SQL query that should be used to discover the schema of a table.
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getSchemaQuery(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getSeed() - Method in interface org.apache.spark.ml.param.shared.HasSeed
-
- getSeed() - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the random seed.
- getSeed() - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Return the random seed
- getSeed() - Method in class org.apache.spark.mllib.clustering.KMeans
-
The random seed for cluster initialization.
- getSeed() - Method in class org.apache.spark.mllib.clustering.LDA
-
Random seed for cluster initialization.
- getSeed() - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
-
Random seed for cluster initialization.
- getSelectionMode() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
- getSelectionThreshold() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
- getSelectorType() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
- getSeq(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of array type as a Scala Seq.
- getSeqOp(boolean, Map<K, Object>, org.apache.spark.util.random.StratifiedSamplingUtils.RandomDataGenerator, Option<Map<K, Object>>) - Static method in class org.apache.spark.util.random.StratifiedSamplingUtils
-
Returns the function used by aggregate to collect sampling statistics for each partition.
- getSequenceCol() - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- getSerializationProxy(Object) - Static method in class org.apache.spark.util.IndylambdaScalaClosures
-
Check if the given reference is a indylambda style Scala closure.
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- getSerializedSize() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest
-
- getSessionConf(SparkSession) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- getShort(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a primitive short.
- getShort(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getShort(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the short type value for rowId
.
- getShorts(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Gets short type values from [rowId, rowId + count)
.
- getShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_bytes_written = 37;
- getShuffleBytesWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_bytes_written = 37;
- getShuffleBytesWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_bytes_written = 37;
- getShuffleChunkCardinality(ShuffleBlockChunkId) - Method in class org.apache.spark.storage.PushBasedFetchHelper
-
Get the number of map blocks in a ShuffleBlockChunk
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_corrupt_merged_block_chunks = 33;
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_corrupt_merged_block_chunks = 33;
- getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_corrupt_merged_block_chunks = 33;
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 53;
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_corrupt_merged_block_chunks = 53;
- getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_corrupt_merged_block_chunks = 53;
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 42;
- getShuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_corrupt_merged_block_chunks = 42;
- getShuffleCorruptMergedBlockChunks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_corrupt_merged_block_chunks = 42;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_fetch_wait_time = 26;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_fetch_wait_time = 26;
- getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_fetch_wait_time = 26;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_fetch_wait_time = 30;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_fetch_wait_time = 30;
- getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_fetch_wait_time = 30;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_fetch_wait_time = 32;
- getShuffleFetchWaitTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_fetch_wait_time = 32;
- getShuffleFetchWaitTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_fetch_wait_time = 32;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_local_blocks_fetched = 25;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_local_blocks_fetched = 25;
- getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_local_blocks_fetched = 25;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_blocks_fetched = 29;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_local_blocks_fetched = 29;
- getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_local_blocks_fetched = 29;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_blocks_fetched = 31;
- getShuffleLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_local_blocks_fetched = 31;
- getShuffleLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_local_blocks_fetched = 31;
- getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_bytes_read = 33;
- getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_local_bytes_read = 33;
- getShuffleLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_local_bytes_read = 33;
- getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_bytes_read = 35;
- getShuffleLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_local_bytes_read = 35;
- getShuffleLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_local_bytes_read = 35;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_fetch_fallback_count = 34;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_fetch_fallback_count = 34;
- getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_fetch_fallback_count = 34;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_fetch_fallback_count = 54;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_fetch_fallback_count = 54;
- getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_fetch_fallback_count = 54;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_fetch_fallback_count = 43;
- getShuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_fetch_fallback_count = 43;
- getShuffleMergedFetchFallbackCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_fetch_fallback_count = 43;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_blocks_fetched = 36;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_local_blocks_fetched = 36;
- getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_local_blocks_fetched = 36;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_blocks_fetched = 56;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_local_blocks_fetched = 56;
- getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_local_blocks_fetched = 56;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_blocks_fetched = 45;
- getShuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_local_blocks_fetched = 45;
- getShuffleMergedLocalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_local_blocks_fetched = 45;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_bytes_read = 40;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_local_bytes_read = 40;
- getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_local_bytes_read = 40;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_bytes_read = 60;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_local_bytes_read = 60;
- getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_local_bytes_read = 60;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_bytes_read = 49;
- getShuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_local_bytes_read = 49;
- getShuffleMergedLocalBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_local_bytes_read = 49;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_chunks_fetched = 38;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_local_chunks_fetched = 38;
- getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_local_chunks_fetched = 38;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_chunks_fetched = 58;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_local_chunks_fetched = 58;
- getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_local_chunks_fetched = 58;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_chunks_fetched = 47;
- getShuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_local_chunks_fetched = 47;
- getShuffleMergedLocalChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_local_chunks_fetched = 47;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_blocks_fetched = 35;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_remote_blocks_fetched = 35;
- getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_remote_blocks_fetched = 35;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 55;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_remote_blocks_fetched = 55;
- getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_remote_blocks_fetched = 55;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 44;
- getShuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_remote_blocks_fetched = 44;
- getShuffleMergedRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_remote_blocks_fetched = 44;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_bytes_read = 39;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_remote_bytes_read = 39;
- getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_remote_bytes_read = 39;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_bytes_read = 59;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_remote_bytes_read = 59;
- getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_remote_bytes_read = 59;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_bytes_read = 48;
- getShuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_remote_bytes_read = 48;
- getShuffleMergedRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_remote_bytes_read = 48;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_chunks_fetched = 37;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_remote_chunks_fetched = 37;
- getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_remote_chunks_fetched = 37;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 57;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_remote_chunks_fetched = 57;
- getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_remote_chunks_fetched = 57;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 46;
- getShuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_remote_chunks_fetched = 46;
- getShuffleMergedRemoteChunksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_remote_chunks_fetched = 46;
- getShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_req_duration = 51;
- getShuffleMergedRemoteReqDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_merged_remote_req_duration = 51;
- getShuffleMergedRemoteReqDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_merged_remote_req_duration = 51;
- getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_reqs_duration = 42;
- getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_merged_remote_reqs_duration = 42;
- getShuffleMergedRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_merged_remote_reqs_duration = 42;
- getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_reqs_duration = 62;
- getShuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_merged_remote_reqs_duration = 62;
- getShuffleMergedRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_merged_remote_reqs_duration = 62;
- getShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 shuffle_mergers_count = 64;
- getShuffleMergersCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int32 shuffle_mergers_count = 64;
- getShuffleMergersCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int32 shuffle_mergers_count = 64;
- getShufflePushMergerLocations(int, int) - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
Get the list of host locations for push based shuffle
- getShufflePushMergerLocations() - Method in class org.apache.spark.ShuffleStatus
-
- GetShufflePushMergerLocations(int, Set<String>) - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetShufflePushMergerLocations
-
- GetShufflePushMergerLocations$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetShufflePushMergerLocations$
-
- getShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDist() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDist() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDistBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDistOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDistOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsDistOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- getShufflePushReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShufflePushReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- getShuffleRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- getShuffleRead(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read = 10;
- getShuffleRead(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read = 10;
- getShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read = 9;
- getShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 shuffle_read = 9;
- getShuffleRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 shuffle_read = 9;
- getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_read_bytes = 22;
- getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_read_bytes = 22;
- getShuffleReadBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_read_bytes = 22;
- getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_bytes = 34;
- getShuffleReadBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_read_bytes = 34;
- getShuffleReadBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_read_bytes = 34;
- getShuffleReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- getShuffleReadCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read = 10;
- getShuffleReadCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read = 10;
- getShuffleReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- getShuffleReadList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read = 10;
- getShuffleReadList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read = 10;
- getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- getShuffleReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read_records = 10;
- getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 shuffle_read_records = 10;
- getShuffleReadRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 shuffle_read_records = 10;
- getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_records = 35;
- getShuffleReadRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_read_records = 35;
- getShuffleReadRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_read_records = 35;
- getShuffleReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_read_records = 11;
- getShuffleReadRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_read_records = 11;
- getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_records_read = 23;
- getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_records_read = 23;
- getShuffleRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_records_read = 23;
- getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_read = 36;
- getShuffleRecordsRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_records_read = 36;
- getShuffleRecordsRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_records_read = 36;
- getShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_written = 39;
- getShuffleRecordsWritten() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_records_written = 39;
- getShuffleRecordsWritten() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_records_written = 39;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_blocks_fetched = 24;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_remote_blocks_fetched = 24;
- getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_remote_blocks_fetched = 24;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_blocks_fetched = 28;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_remote_blocks_fetched = 28;
- getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_remote_blocks_fetched = 28;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_blocks_fetched = 30;
- getShuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_remote_blocks_fetched = 30;
- getShuffleRemoteBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_remote_blocks_fetched = 30;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read = 27;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_remote_bytes_read = 27;
- getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_remote_bytes_read = 27;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read = 31;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_remote_bytes_read = 31;
- getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_remote_bytes_read = 31;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read = 33;
- getShuffleRemoteBytesRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_remote_bytes_read = 33;
- getShuffleRemoteBytesRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_remote_bytes_read = 33;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read_to_disk = 28;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_remote_bytes_read_to_disk = 28;
- getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_remote_bytes_read_to_disk = 28;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 32;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_remote_bytes_read_to_disk = 32;
- getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_remote_bytes_read_to_disk = 32;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 34;
- getShuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_remote_bytes_read_to_disk = 34;
- getShuffleRemoteBytesReadToDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_remote_bytes_read_to_disk = 34;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_reqs_duration = 41;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_remote_reqs_duration = 41;
- getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_remote_reqs_duration = 41;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_reqs_duration = 61;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_remote_reqs_duration = 61;
- getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_remote_reqs_duration = 61;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_reqs_duration = 50;
- getShuffleRemoteReqsDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_remote_reqs_duration = 50;
- getShuffleRemoteReqsDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_remote_reqs_duration = 50;
- getShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_total_blocks_fetched = 29;
- getShuffleTotalBlocksFetched() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_total_blocks_fetched = 29;
- getShuffleTotalBlocksFetched() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_total_blocks_fetched = 29;
- getShuffleWrite(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- getShuffleWrite(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write = 12;
- getShuffleWrite(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write = 12;
- getShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write = 11;
- getShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 shuffle_write = 11;
- getShuffleWrite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 shuffle_write = 11;
- getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_bytes = 30;
- getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_write_bytes = 30;
- getShuffleWriteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_write_bytes = 30;
- getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_bytes = 36;
- getShuffleWriteBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_write_bytes = 36;
- getShuffleWriteBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_write_bytes = 36;
- getShuffleWriteCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- getShuffleWriteCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write = 12;
- getShuffleWriteCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write = 12;
- getShuffleWriteList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- getShuffleWriteList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write = 12;
- getShuffleWriteList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write = 12;
- getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributionsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskMetricsOrBuilder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_records = 31;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_write_records = 31;
- getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_write_records = 31;
- getShuffleWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write_records = 12;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 shuffle_write_records = 12;
- getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 shuffle_write_records = 12;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_records = 38;
- getShuffleWriteRecords() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_write_records = 38;
- getShuffleWriteRecords() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_write_records = 38;
- getShuffleWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double shuffle_write_records = 13;
- getShuffleWriteRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double shuffle_write_records = 13;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_time = 32;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
double shuffle_write_time = 32;
- getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
double shuffle_write_time = 32;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_time = 37;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 shuffle_write_time = 37;
- getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 shuffle_write_time = 37;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_write_time = 38;
- getShuffleWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 shuffle_write_time = 38;
- getShuffleWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 shuffle_write_time = 38;
- getSimpleMessage() - Method in exception org.apache.spark.sql.AnalysisException
-
- getSimpleName(Class<?>) - Static method in class org.apache.spark.util.Utils
-
Safer than Class obj's getSimpleName which may throw Malformed class name error in scala.
- getSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSink() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSink() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSinkBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSinkOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSinkOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSinkOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- getSize() - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
group getParam
- getSizeAsBytes(String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as bytes; throws a NoSuchElementException if it's not set.
- getSizeAsBytes(String, String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as bytes, falling back to a default if not set.
- getSizeAsBytes(String, long) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as bytes, falling back to a default if not set.
- getSizeAsGb(String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Gibibytes; throws a NoSuchElementException if it's not set.
- getSizeAsGb(String, String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Gibibytes, falling back to a default if not set.
- getSizeAsKb(String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Kibibytes; throws a NoSuchElementException if it's not set.
- getSizeAsKb(String, String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Kibibytes, falling back to a default if not set.
- getSizeAsMb(String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Mebibytes; throws a NoSuchElementException if it's not set.
- getSizeAsMb(String, String) - Method in class org.apache.spark.SparkConf
-
Get a size parameter as Mebibytes, falling back to a default if not set.
- getSizeForBlock(int) - Method in interface org.apache.spark.scheduler.MapStatus
-
Estimated size for the reduce block, in bytes.
- getSizeInBytes() - Method in interface org.apache.spark.ml.linalg.Matrix
-
Gets the current size in bytes of this `Matrix`.
- getSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- getSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
repeated int32 skipped_stages = 2;
- getSkippedStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
repeated int32 skipped_stages = 2;
- getSkippedStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- getSkippedStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
repeated int32 skipped_stages = 2;
- getSkippedStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
repeated int32 skipped_stages = 2;
- getSkippedStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- getSkippedStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
repeated int32 skipped_stages = 2;
- getSkippedStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
repeated int32 skipped_stages = 2;
- getSlotDescs() - Method in class org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getSmoothing() - Method in interface org.apache.spark.ml.classification.NaiveBayesParams
-
- getSolver() - Method in interface org.apache.spark.ml.param.shared.HasSolver
-
- getSortedTaskSetQueue() - Method in interface org.apache.spark.scheduler.Schedulable
-
- getSources(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSources(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSources(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSourcesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- getSparkClassLoader() - Method in interface org.apache.spark.util.SparkClassUtils
-
- getSparkClassLoader() - Static method in class org.apache.spark.util.Utils
-
- getSparkHome() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Get Spark's home location from either a value set through the constructor,
or the spark.home Java property, or the SPARK_HOME environment variable
(in that order of preference).
- getSparkOrYarnConfig(SparkConf, String, String) - Static method in class org.apache.spark.util.Utils
-
Return the value of a config either through the SparkConf or the Hadoop configuration.
- getSparkProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- getSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string spark_user = 6;
- getSparkUser() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string spark_user = 6;
- getSparkUser() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string spark_user = 6;
- getSparkUserBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string spark_user = 6;
- getSparkUserBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
optional string spark_user = 6;
- getSparkUserBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
optional string spark_user = 6;
- getSparseSizeInBytes(boolean) - Method in interface org.apache.spark.ml.linalg.Matrix
-
Gets the size of the minimal sparse representation of this `Matrix`.
- getSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummary() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummary() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummaryBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummaryOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummaryOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculationSummaryOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
bool speculative = 12;
- getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
bool speculative = 12;
- getSpeculative() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
bool speculative = 12;
- getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool speculative = 12;
- getSpeculative() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
bool speculative = 12;
- getSpeculative() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
bool speculative = 12;
- getSplit() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- getSplits() - Method in class org.apache.spark.ml.feature.Bucketizer
-
- getSplitsArray() - Method in class org.apache.spark.ml.feature.Bucketizer
-
- getSql() - Method in class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
-
Returns the SQL string (Spark SQL dialect) of the default value expression.
- getSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
optional int64 sql_execution_id = 3;
- getSqlExecutionId() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
optional int64 sql_execution_id = 3;
- getSqlExecutionId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataWrapperOrBuilder
-
optional int64 sql_execution_id = 3;
- getSqlState(String) - Method in class org.apache.spark.ErrorClassesJsonReader
-
- getSqlState() - Method in interface org.apache.spark.SparkThrowable
-
- getSqlState(String) - Static method in class org.apache.spark.SparkThrowableHelper
-
- getSrcCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
-
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
-
int32 stage_attempt_id = 2;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 stage_attempt_id = 41;
- getStageAttemptId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int32 stage_attempt_id = 41;
- getStageAttemptId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int32 stage_attempt_id = 41;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
int64 stage_id = 1;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
int64 stage_id = 1;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapperOrBuilder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
int64 stage_id = 1;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapperOrBuilder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
int64 stage_id = 1;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapperOrBuilder
-
int64 stage_id = 1;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 stage_id = 2;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
int64 stage_id = 2;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
int64 stage_id = 2;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 stage_id = 40;
- getStageId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 stage_id = 40;
- getStageId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 stage_id = 40;
- getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated int64 stage_ids = 6;
- getStageIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated int64 stage_ids = 6;
- getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- getStageIds(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
repeated int64 stage_ids = 2;
- getStageIds(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
-
repeated int64 stage_ids = 2;
- getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated int64 stage_ids = 6;
- getStageIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated int64 stage_ids = 6;
- getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- getStageIdsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
repeated int64 stage_ids = 2;
- getStageIdsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
-
repeated int64 stage_ids = 2;
- getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
repeated int64 stage_ids = 6;
- getStageIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
repeated int64 stage_ids = 6;
- getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- getStageIdsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
repeated int64 stage_ids = 2;
- getStageIdsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PoolDataOrBuilder
-
repeated int64 stage_ids = 2;
- getStageInfo(int) - Method in class org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns stage information, or null
if the stage info could not be found or was
garbage collected.
- getStageInfo(int) - Method in class org.apache.spark.SparkStatusTracker
-
Returns stage information, or None
if the stage info could not be found or was
garbage collected.
- getStagePath(String, int, int, String) - Method in class org.apache.spark.ml.Pipeline.SharedReadWrite$
-
Get path for saving the given stage.
- getStages() - Method in class org.apache.spark.ml.Pipeline
-
- getStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- getStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated int64 stages = 12;
- getStages(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated int64 stages = 12;
- getStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- getStagesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated int64 stages = 12;
- getStagesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated int64 stages = 12;
- getStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- getStagesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
repeated int64 stages = 12;
- getStagesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
repeated int64 stages = 12;
- getStandardization() - Method in interface org.apache.spark.ml.param.shared.HasStandardization
-
- getStartOffset() - Static method in class org.apache.spark.rdd.InputFileBlockHolder
-
Returns the starting offset of the block currently being read, or -1 if it is unknown.
- getStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string start_offset = 2;
- getStartOffset() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string start_offset = 2;
- getStartOffset() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string start_offset = 2;
- getStartOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string start_offset = 2;
- getStartOffsetBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
optional string start_offset = 2;
- getStartOffsetBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder
-
optional string start_offset = 2;
- getStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 start_time = 2;
- getStartTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
int64 start_time = 2;
- getStartTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfoOrBuilder
-
int64 start_time = 2;
- getStartTimeEpoch() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
int64 start_timestamp = 6;
- getStartTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
int64 start_timestamp = 6;
- getStartTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryDataOrBuilder
-
int64 start_timestamp = 6;
- getState() - Method in interface org.apache.spark.launcher.SparkAppHandle
-
Returns the current application state.
- getState() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
:: DeveloperApi ::
- getState() - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
:: DeveloperApi ::
- getStatement() - Method in class org.apache.spark.ml.feature.SQLTransformer
-
- getStateOperators(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperators(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperators(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStateOperatorsOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string status = 10;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string status = 10;
- getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string status = 10;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string status = 10;
- getStatus() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string status = 10;
- getStatus() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string status = 10;
- getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string status = 10;
- getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string status = 10;
- getStatusBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string status = 10;
- getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string status = 10;
- getStatusBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string status = 10;
- getStatusBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string status = 10;
- getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatusValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStatusValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStatusValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- getStderr(Process, long) - Static method in class org.apache.spark.util.Utils
-
Return the stderr of a process after waiting for the process to terminate.
- getStepSize() - Method in interface org.apache.spark.ml.param.shared.HasStepSize
-
- getStopWords() - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- getStorageLevel() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
- getStorageLevel() - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
- getStorageLevel() - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
- getStorageLevel() - Method in class org.apache.spark.rdd.RDD
-
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string storage_level = 2;
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
optional string storage_level = 2;
- getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
optional string storage_level = 2;
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string storage_level = 5;
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
optional string storage_level = 5;
- getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
optional string storage_level = 5;
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string storage_level = 4;
- getStorageLevel() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string storage_level = 4;
- getStorageLevel() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string storage_level = 4;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string storage_level = 2;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
optional string storage_level = 2;
- getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfoOrBuilder
-
optional string storage_level = 2;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string storage_level = 5;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
optional string storage_level = 5;
- getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoOrBuilder
-
optional string storage_level = 5;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string storage_level = 4;
- getStorageLevelBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
optional string storage_level = 4;
- getStorageLevelBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
optional string storage_level = 4;
- GetStorageStatus$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.GetStorageStatus$
-
- getStrategy() - Method in interface org.apache.spark.ml.feature.ImputerParams
-
- getString(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i as a String object.
- getString(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a String.
- getStringArray(String) - Method in class org.apache.spark.sql.types.Metadata
-
Gets a String array.
- getStringField(boolean, Function0<String>) - Static method in class org.apache.spark.status.protobuf.Utils
-
- getStringIndexerOrderType() - Method in interface org.apache.spark.ml.feature.RFormulaBase
-
- getStringOrderType() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
-
- getStruct(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of struct type as a
Row
object.
- getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getStruct(int, int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getStruct(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the struct type value for rowId
.
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 submission_time = 4;
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
optional int64 submission_time = 4;
- getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder
-
optional int64 submission_time = 4;
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 submission_time = 8;
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
int64 submission_time = 8;
- getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder
-
int64 submission_time = 8;
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 submission_time = 10;
- getSubmissionTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional int64 submission_time = 10;
- getSubmissionTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional int64 submission_time = 10;
- getSubsamplingRate() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getSubsamplingRate() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
-
- getSubsamplingRate() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getSucceededTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- getSucceededTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double succeeded_tasks = 4;
- getSucceededTasks(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double succeeded_tasks = 4;
- getSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 succeeded_tasks = 3;
- getSucceededTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int32 succeeded_tasks = 3;
- getSucceededTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int32 succeeded_tasks = 3;
- getSucceededTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- getSucceededTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double succeeded_tasks = 4;
- getSucceededTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double succeeded_tasks = 4;
- getSucceededTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- getSucceededTasksList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double succeeded_tasks = 4;
- getSucceededTasksList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double succeeded_tasks = 4;
- getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- getSummary(SQLQueryContext) - Method in interface org.apache.spark.sql.errors.DataTypeErrorsBase
-
- getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- getSummary(SQLQueryContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- getSystemProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemProperties(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemProperties(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemProperties() - Static method in class org.apache.spark.util.Utils
-
Returns the system properties map that is thread-safe to iterator over.
- getSystemPropertiesBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilder(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilder(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilderList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getSystemPropertiesOrBuilderList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoOrBuilder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- getTable(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Get the table or view with the specified name.
- getTable(String, String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Get the table or view with the specified name in the specified database under the Hive
Metastore.
- getTable(CatalogPlugin, Identifier, Option<TimeTravelSpec>) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
- getTable(StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.TableProvider
-
Return a
Table
instance with the specified table schema, partitioning and properties
to do read/write.
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTableCommentQuery(String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getTableCommentQuery(String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getTableExistsQuery(String) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTableExistsQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Get the SQL query that should be used to find if the given table exists.
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getTableExistsQuery(String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getTableNames(SparkSession, String) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- getTableParameters(HttpServletRequest, String, String) - Method in interface org.apache.spark.ui.PagedTable
-
Returns parameter of this table.
- getTableProviderCatalog(SupportsCatalogOptions, CatalogManager, CaseInsensitiveStringMap) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTableSample(TableSampleInfo) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getTableSample(TableSampleInfo) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getTablesByTypeUnsupportedByHiveVersionError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- getTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 task_count = 4;
- getTaskCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
int64 task_count = 4;
- getTaskCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.CachedQuantileOrBuilder
-
int64 task_count = 4;
- getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 task_id = 1;
- getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
int64 task_id = 1;
- getTaskId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
int64 task_id = 1;
- getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 task_id = 1;
- getTaskId() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
int64 task_id = 1;
- getTaskId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
int64 task_id = 1;
- getTaskInfos() - Method in class org.apache.spark.BarrierTaskContext
-
:: Experimental ::
Returns
BarrierTaskInfo
for all tasks in this barrier stage, ordered by partition ID.
- getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string task_locality = 11;
- getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string task_locality = 11;
- getTaskLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string task_locality = 11;
- getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string task_locality = 11;
- getTaskLocality() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string task_locality = 11;
- getTaskLocality() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
optional string task_locality = 11;
- getTaskLocalityBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapperOrBuilder
-
optional string task_locality = 11;
- getTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetrics() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetrics() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetricsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributions() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributions() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributionsBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributionsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsDistributionsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- getTaskMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetricsOrBuilder() - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskMetricsOrBuilder() - Method in interface org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- getTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
Deprecated.
- getTaskResources() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
Deprecated.
- getTaskResources() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
Deprecated.
- getTaskResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- getTaskResourcesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
- getTaskResourcesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrDefault(String, StoreTypes.TaskResourceRequest) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrThrow(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTaskResourcesOrThrow(String) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder
-
map<string, .org.apache.spark.status.protobuf.TaskResourceRequest> task_resources = 3;
- getTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
Deprecated.
- getTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
Deprecated.
- getTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
Deprecated.
- getTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- getTasksCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- getTasksCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksMap() - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksMap() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrDefault(long, StoreTypes.TaskData) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrDefault(long, StoreTypes.TaskData) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrDefault(long, StoreTypes.TaskData) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrThrow(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTasksOrThrow(long) - Method in interface org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder
-
map<int64, .org.apache.spark.status.protobuf.TaskData> tasks = 45;
- getTaskTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- getTaskTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double task_time = 2;
- getTaskTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double task_time = 2;
- getTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 task_time = 1;
- getTaskTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
int64 task_time = 1;
- getTaskTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryOrBuilder
-
int64 task_time = 1;
- getTaskTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- getTaskTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double task_time = 2;
- getTaskTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double task_time = 2;
- getTaskTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- getTaskTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
repeated double task_time = 2;
- getTaskTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributionsOrBuilder
-
repeated double task_time = 2;
- getTau0() - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
A (positive) learning parameter that downweights early iterations.
- getThreadDump() - Static method in class org.apache.spark.util.Utils
-
Return a thread dump of all threads' stacktraces.
- getThreadDumpForThread(long) - Static method in class org.apache.spark.util.Utils
-
- getThreshold() - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- getThreshold() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- getThreshold() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
Get threshold for binary classification.
- getThreshold() - Method in interface org.apache.spark.ml.param.shared.HasThreshold
-
- getThreshold() - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
-
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
- getThreshold() - Method in class org.apache.spark.mllib.classification.SVMModel
-
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
- getThresholds() - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- getThresholds() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- getThresholds() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
Get thresholds for binary or multiclass classification.
- getThresholds() - Method in interface org.apache.spark.ml.param.shared.HasThresholds
-
- getThroughOrigin() - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getTimeAsMs(String) - Method in class org.apache.spark.SparkConf
-
Get a time parameter as milliseconds; throws a NoSuchElementException if it's not set.
- getTimeAsMs(String, String) - Method in class org.apache.spark.SparkConf
-
Get a time parameter as milliseconds, falling back to a default if not set.
- getTimeAsSeconds(String) - Method in class org.apache.spark.SparkConf
-
Get a time parameter as seconds; throws a NoSuchElementException if it's not set.
- getTimeAsSeconds(String, String) - Method in class org.apache.spark.SparkConf
-
Get a time parameter as seconds, falling back to a default if not set.
- getTimeMillis() - Method in interface org.apache.spark.util.Clock
-
- getTimeoutTimestampMs() - Method in interface org.apache.spark.sql.streaming.TestGroupState
-
Returns the timestamp if setTimeoutTimestamp()
is called.
- getTimer(L) - Method in interface org.apache.spark.util.ListenerBus
-
Returns a CodaHale metrics Timer for measuring the listener's event processing time.
- getTimestamp(int) - Method in interface org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.sql.Timestamp.
- getTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string timestamp = 4;
- getTimestamp() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string timestamp = 4;
- getTimestamp() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string timestamp = 4;
- getTimestamp() - Method in class org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- getTimestampBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string timestamp = 4;
- getTimestampBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
optional string timestamp = 4;
- getTimestampBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder
-
optional string timestamp = 4;
- getTimeZoneOffset() - Static method in class org.apache.spark.ui.UIUtils
-
- GETTING_RESULT_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- GETTING_RESULT_TIME() - Static method in class org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- GETTING_RESULT_TIME() - Static method in class org.apache.spark.ui.ToolTips
-
- GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- GETTING_RESULT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- gettingResult() - Method in class org.apache.spark.scheduler.TaskInfo
-
- gettingResultTime() - Method in class org.apache.spark.scheduler.TaskInfo
-
The time when the task started remotely getting the result.
- gettingResultTime() - Method in class org.apache.spark.status.api.v1.TaskData
-
- gettingResultTime() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
-
- gettingResultTime(TaskData) - Static method in class org.apache.spark.status.AppStatusUtils
-
- gettingResultTime(long, long, long) - Static method in class org.apache.spark.status.AppStatusUtils
-
- getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 to_id = 2;
- getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge
-
int32 to_id = 2;
- getToId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdgeOrBuilder
-
int32 to_id = 2;
- getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 to_id = 2;
- getToId() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge
-
int64 to_id = 2;
- getToId() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdgeOrBuilder
-
int64 to_id = 2;
- getTokenJaasParams(KafkaTokenClusterConf) - Static method in class org.apache.spark.kafka010.KafkaTokenUtil
-
- getTol() - Method in interface org.apache.spark.ml.param.shared.HasTol
-
- getToLowercase() - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- getTopicConcentration() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getTopicConcentration() - Method in class org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics'
distributions over terms.
- getTopicDistributionCol() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
- getTopologyForHost(String) - Method in class org.apache.spark.storage.DefaultTopologyMapper
-
- getTopologyForHost(String) - Method in class org.apache.spark.storage.FileBasedTopologyMapper
-
- getTopologyForHost(String) - Method in class org.apache.spark.storage.TopologyMapper
-
Gets the topology information given the host name
- getTotalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetched(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetched(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
repeated double total_blocks_fetched = 8;
- getTotalBlocksFetchedList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributionsOrBuilder
-
repeated double total_blocks_fetched = 8;
- getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_cores = 7;
- getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 total_cores = 7;
- getTotalCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 total_cores = 7;
- getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int32 total_cores = 4;
- getTotalCores() - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary
-
int32 total_cores = 4;
- getTotalCores() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder
-
int32 total_cores = 4;
- getTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_duration = 13;
- getTotalDuration() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 total_duration = 13;
- getTotalDuration() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 total_duration = 13;
- getTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_gc_time = 14;
- getTotalGcTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 total_gc_time = 14;
- getTotalGcTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 total_gc_time = 14;
- getTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_input_bytes = 15;
- getTotalInputBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 total_input_bytes = 15;
- getTotalInputBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 total_input_bytes = 15;
- getTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_off_heap_storage_memory = 4;
- getTotalOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
int64 total_off_heap_storage_memory = 4;
- getTotalOffHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
-
int64 total_off_heap_storage_memory = 4;
- getTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_on_heap_storage_memory = 3;
- getTotalOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
int64 total_on_heap_storage_memory = 3;
- getTotalOnHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
-
int64 total_on_heap_storage_memory = 3;
- getTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_read = 16;
- getTotalShuffleRead() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 total_shuffle_read = 16;
- getTotalShuffleRead() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 total_shuffle_read = 16;
- getTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_write = 17;
- getTotalShuffleWrite() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int64 total_shuffle_write = 17;
- getTotalShuffleWrite() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int64 total_shuffle_write = 17;
- getTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_tasks = 12;
- getTotalTasks() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary
-
int32 total_tasks = 12;
- getTotalTasks() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder
-
int32 total_tasks = 12;
- getTrainRatio() - Method in interface org.apache.spark.ml.tuning.TrainValidationSplitParams
-
- getTreeStrategy() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getTruncateQuery(String, Option<Object>) - Method in class org.apache.spark.sql.jdbc.AggregatedDialect
-
The SQL query used to truncate a table.
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTruncateQuery(String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
The SQL query that should be used to truncate a table.
- getTruncateQuery(String, Option<Object>) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
The SQL query that should be used to truncate a table.
- getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTruncateQuery(String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
The SQL query used to truncate a table.
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
The SQL query used to truncate a table.
- getTruncateQuery(String, Option<Object>) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
The SQL query used to truncate a table.
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getTruncateQuery$default$2() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getUDTFor(String) - Static method in class org.apache.spark.sql.types.UDTRegistration
-
Returns the Class of UserDefinedType for the name of a given user class.
- getUidMap(Params) - Static method in class org.apache.spark.ml.util.MetaAlgorithmReadWrite
-
Examine the given estimator (which may be a compound estimator) and extract a mapping
from UIDs to corresponding Params
instances.
- getUiRoot(ServletContext) - Static method in class org.apache.spark.status.api.v1.UIRootFromServletContext
-
- getUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string update = 3;
- getUpdate() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string update = 3;
- getUpdate() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string update = 3;
- getUpdateBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string update = 3;
- getUpdateBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string update = 3;
- getUpdateBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string update = 3;
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getUpdateColumnNullabilityQuery(String, String, boolean) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- getUpdateColumnTypeQuery(String, String, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- getUpdateColumnTypeQuery(String, String, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- getUpper() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
-
- getUpperBound(double, long, double) - Static method in class org.apache.spark.util.random.BinomialBounds
-
Returns a threshold p
such that if we conduct n Bernoulli trials with success rate = p
,
it is very unlikely to have less than fraction * n
successes.
- getUpperBound(double) - Static method in class org.apache.spark.util.random.PoissonBounds
-
Returns a lambda such that Pr[X < s] is very small, where X ~ Pois(lambda).
- getUpperBoundsOnCoefficients() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
- getUpperBoundsOnIntercepts() - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
- getUsedBins() - Method in class org.apache.spark.sql.util.NumericHistogram
-
Returns the number of bins currently being used by the histogram.
- getUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_disk = 6;
- getUseDisk() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
bool use_disk = 6;
- getUseDisk() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
bool use_disk = 6;
- getUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_off_heap_storage_memory = 2;
- getUsedOffHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
int64 used_off_heap_storage_memory = 2;
- getUsedOffHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
-
int64 used_off_heap_storage_memory = 2;
- getUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_on_heap_storage_memory = 1;
- getUsedOnHeapStorageMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics
-
int64 used_on_heap_storage_memory = 1;
- getUsedOnHeapStorageMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.MemoryMetricsOrBuilder
-
int64 used_on_heap_storage_memory = 1;
- getUsedTimeNs(long) - Static method in class org.apache.spark.util.Utils
-
Return the string to tell how long has passed in milliseconds.
- getUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_memory = 5;
- getUseMemory() - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
bool use_memory = 5;
- getUseMemory() - Method in interface org.apache.spark.status.protobuf.StoreTypes.StreamBlockDataOrBuilder
-
bool use_memory = 5;
- getUseNodeIdCache() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- getUserCol() - Method in interface org.apache.spark.ml.recommendation.ALSModelParams
-
- getUserJars(SparkConf) - Static method in class org.apache.spark.util.Utils
-
Return the jar files pointed by the "spark.jars" property.
- getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- getUTF8String(int) - Method in class org.apache.spark.sql.vectorized.ColumnVector
-
Returns the string type value for rowId
.
- getValidationIndicatorCol() - Method in interface org.apache.spark.ml.param.shared.HasValidationIndicatorCol
-
- getValidationTol() - Method in interface org.apache.spark.ml.tree.GBTParams
-
- getValidationTol() - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getValue(int) - Method in class org.apache.spark.ml.attribute.NominalAttribute
-
Gets a value given its index.
- getValue() - Method in class org.apache.spark.mllib.stat.test.BinarySample
-
- getValue() - Method in class org.apache.spark.sql.connector.catalog.ColumnDefaultValue
-
Returns the default value literal.
- getValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string value = 4;
- getValue() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string value = 4;
- getValue() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string value = 4;
- getValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value1 = 1;
- getValue1() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
optional string value1 = 1;
- getValue1() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
-
optional string value1 = 1;
- getValue1Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value1 = 1;
- getValue1Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
optional string value1 = 1;
- getValue1Bytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
-
optional string value1 = 1;
- getValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value2 = 2;
- getValue2() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
optional string value2 = 2;
- getValue2() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
-
optional string value2 = 2;
- getValue2Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value2 = 2;
- getValue2Bytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings
-
optional string value2 = 2;
- getValue2Bytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.PairStringsOrBuilder
-
optional string value2 = 2;
- getValueBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string value = 4;
- getValueBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo
-
optional string value = 4;
- getValueBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.AccumulableInfoOrBuilder
-
optional string value = 4;
- getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel
-
- getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus
-
- getValueDescriptor() - Method in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
- getValuesMap(Seq<String>) - Method in interface org.apache.spark.sql.Row
-
Returns a Map consisting of names and values for the requested fieldNames
For primitive types if value is null it returns 'zero value' specific for primitive
i.e.
- getValueVector() - Method in class org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getVarianceCol() - Method in interface org.apache.spark.ml.param.shared.HasVarianceCol
-
- getVariancePower() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
- getVarianceThreshold() - Method in interface org.apache.spark.ml.feature.VarianceThresholdSelectorParams
-
- getVectors() - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- getVectors() - Method in class org.apache.spark.mllib.feature.Word2VecModel
-
Returns a map of words to their vector representations.
- getVectorSize() - Method in interface org.apache.spark.ml.feature.Word2VecBase
-
- getVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string vendor = 4;
- getVendor() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string vendor = 4;
- getVendor() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string vendor = 4;
- getVendorBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string vendor = 4;
- getVendorBytes() - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest
-
optional string vendor = 4;
- getVendorBytes() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequestOrBuilder
-
optional string vendor = 4;
- getVocabSize() - Method in interface org.apache.spark.ml.feature.CountVectorizerParams
-
- getWeightCol() - Method in interface org.apache.spark.ml.param.shared.HasWeightCol
-
- getWidth(Row) - Static method in class org.apache.spark.ml.image.ImageSchema
-
Gets the width of the image
- getWindowSize() - Method in interface org.apache.spark.ml.feature.Word2VecBase
-
- getWithCentering() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
-
- getWithMean() - Method in interface org.apache.spark.ml.feature.StandardScalerParams
-
- getWithScaling() - Method in interface org.apache.spark.ml.feature.RobustScalerParams
-
- getWithStd() - Method in interface org.apache.spark.ml.feature.StandardScalerParams
-
- getWrapperCase() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- getWrapperCase() - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper
-
- getWrapperCase() - Method in interface org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapperOrBuilder
-
- getWriteBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- getWriteBytes(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_bytes = 1;
- getWriteBytes(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_bytes = 1;
- getWriteBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- getWriteBytesCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_bytes = 1;
- getWriteBytesCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_bytes = 1;
- getWriteBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- getWriteBytesList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_bytes = 1;
- getWriteBytesList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_bytes = 1;
- getWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- getWriteRecords(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_records = 2;
- getWriteRecords(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_records = 2;
- getWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- getWriteRecordsCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_records = 2;
- getWriteRecordsCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_records = 2;
- getWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- getWriteRecordsList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_records = 2;
- getWriteRecordsList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_records = 2;
- getWriteTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- getWriteTime(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_time = 3;
- getWriteTime(int) - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_time = 3;
- getWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 write_time = 2;
- getWriteTime() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics
-
int64 write_time = 2;
- getWriteTime() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricsOrBuilder
-
int64 write_time = 2;
- getWriteTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- getWriteTimeCount() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_time = 3;
- getWriteTimeCount() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_time = 3;
- getWriteTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- getWriteTimeList() - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions
-
repeated double write_time = 3;
- getWriteTimeList() - Method in interface org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributionsOrBuilder
-
repeated double write_time = 3;
- Gini - Class in org.apache.spark.mllib.tree.impurity
-
Class for calculating the Gini impurity
(http://en.wikipedia.org/wiki/Decision_tree_learning#Gini_impurity)
during multiclass classification.
- Gini() - Constructor for class org.apache.spark.mllib.tree.impurity.Gini
-
- GLMClassificationModel - Class in org.apache.spark.mllib.classification.impl
-
Helper class for import/export of GLM classification models.
- GLMClassificationModel() - Constructor for class org.apache.spark.mllib.classification.impl.GLMClassificationModel
-
- GLMClassificationModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.classification.impl
-
- GLMClassificationModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.classification.impl
-
Model data for import/export
- GLMClassificationModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.classification.impl
-
- GLMRegressionModel - Class in org.apache.spark.mllib.regression.impl
-
Helper methods for import/export of GLM regression models.
- GLMRegressionModel() - Constructor for class org.apache.spark.mllib.regression.impl.GLMRegressionModel
-
- GLMRegressionModel.SaveLoadV1_0$ - Class in org.apache.spark.mllib.regression.impl
-
- GLMRegressionModel.SaveLoadV1_0$.Data - Class in org.apache.spark.mllib.regression.impl
-
Model data for model import/export
- GLMRegressionModel.SaveLoadV1_0$.Data$ - Class in org.apache.spark.mllib.regression.impl
-
- glom() - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by coalescing all elements within each partition into an array.
- glom() - Method in class org.apache.spark.rdd.RDD
-
Return an RDD created by coalescing all elements within each partition into an array.
- glom() - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying glom() to each RDD of
this DStream.
- glom() - Method in class org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying glom() to each RDD of
this DStream.
- goButtonFormPath() - Method in interface org.apache.spark.ui.PagedTable
-
Returns the submission path for the "go to page #" form.
- goodnessOfFit() - Method in class org.apache.spark.mllib.stat.test.ChiSqTest.NullHypothesis$
-
- GPU() - Static method in class org.apache.spark.resource.ResourceUtils
-
- grad(DenseMatrix<Object>, DenseMatrix<Object>, DenseVector<Object>) - Method in interface org.apache.spark.ml.ann.LayerModel
-
Computes the gradient.
- grad() - Method in class org.apache.spark.mllib.optimization.NNLS.Workspace
-
- gradient() - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
-
The current weighted averaged gradient.
- Gradient - Class in org.apache.spark.mllib.optimization
-
Class used to compute the gradient for a loss function, given a single data point.
- Gradient() - Constructor for class org.apache.spark.mllib.optimization.Gradient
-
- gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.AbsoluteError
-
Method to calculate the gradients for the gradient boosting calculation for least
absolute error calculation.
- gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.LogLoss
-
Method to calculate the loss gradients for the gradient boosting calculation for binary
classification
The gradient with respect to F(x) is: - 4 y / (1 + exp(2 y F(x)))
- gradient(double, double) - Method in interface org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate the gradients for the gradient boosting calculation.
- gradient(double, double) - Static method in class org.apache.spark.mllib.tree.loss.SquaredError
-
Method to calculate the gradients for the gradient boosting calculation for least
squares error calculation.
- GradientBoostedTrees - Class in org.apache.spark.ml.tree.impl
-
- GradientBoostedTrees() - Constructor for class org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
- GradientBoostedTrees - Class in org.apache.spark.mllib.tree
-
- GradientBoostedTrees(BoostingStrategy) - Constructor for class org.apache.spark.mllib.tree.GradientBoostedTrees
-
- GradientBoostedTreesModel - Class in org.apache.spark.mllib.tree.model
-
Represents a gradient boosted trees model.
- GradientBoostedTreesModel(Enumeration.Value, DecisionTreeModel[], double[]) - Constructor for class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- GradientDescent - Class in org.apache.spark.mllib.optimization
-
Class used to solve an optimization problem using Gradient Descent.
- gradientSumArray() - Method in interface org.apache.spark.ml.optim.aggregator.DifferentiableLossAggregator
-
Array of gradient values that are mutated when new instances are added to the aggregator.
- Graph<VD,ED> - Class in org.apache.spark.graphx
-
The Graph abstractly represents a graph with arbitrary objects
associated with vertices and edges.
- GraphGenerators - Class in org.apache.spark.graphx.util
-
A collection of graph generating functions.
- GraphGenerators() - Constructor for class org.apache.spark.graphx.util.GraphGenerators
-
- GraphImpl<VD,ED> - Class in org.apache.spark.graphx.impl
-
An implementation of
Graph
to support computation on graphs.
- graphiteSinkInvalidProtocolError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- graphiteSinkPropertyMissingError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- GraphLoader - Class in org.apache.spark.graphx
-
Provides utilities for loading
Graph
s from files.
- GraphLoader() - Constructor for class org.apache.spark.graphx.GraphLoader
-
- GraphOps<VD,ED> - Class in org.apache.spark.graphx
-
Contains additional functionality for
Graph
.
- GraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - Constructor for class org.apache.spark.graphx.GraphOps
-
- graphToGraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - Static method in class org.apache.spark.graphx.Graph
-
Implicitly extracts the
GraphOps
member from a graph.
- GraphXUtils - Class in org.apache.spark.graphx
-
- GraphXUtils() - Constructor for class org.apache.spark.graphx.GraphXUtils
-
- greater(Duration) - Method in class org.apache.spark.streaming.Duration
-
- greater(Time) - Method in class org.apache.spark.streaming.Time
-
- greaterEq(Duration) - Method in class org.apache.spark.streaming.Duration
-
- greaterEq(Time) - Method in class org.apache.spark.streaming.Time
-
- GreaterThan - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff the attribute evaluates to a value
greater than value
.
- GreaterThan(String, Object) - Constructor for class org.apache.spark.sql.sources.GreaterThan
-
- GreaterThanOrEqual - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff the attribute evaluates to a value
greater than or equal to value
.
- GreaterThanOrEqual(String, Object) - Constructor for class org.apache.spark.sql.sources.GreaterThanOrEqual
-
- greatest(Column...) - Static method in class org.apache.spark.sql.functions
-
Returns the greatest value of the list of values, skipping null values.
- greatest(String, String...) - Static method in class org.apache.spark.sql.functions
-
Returns the greatest value of the list of column names, skipping null values.
- greatest(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Returns the greatest value of the list of values, skipping null values.
- greatest(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
-
Returns the greatest value of the list of column names, skipping null values.
- gridGraph(SparkContext, int, int) - Static method in class org.apache.spark.graphx.util.GraphGenerators
-
Create rows
by cols
grid graph with each vertex connected to its
row+1 and col+1 neighbors.
- groupArr() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
- groupBy(Function<T, U>) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return an RDD of grouped elements.
- groupBy(Function<T, U>, int) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Return an RDD of grouped elements.
- groupBy(Function1<T, K>, ClassTag<K>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD of grouped items.
- groupBy(Function1<T, K>, int, ClassTag<K>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD of grouped elements.
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD of grouped items.
- groupBy(Column...) - Method in class org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so we can run aggregation on them.
- groupBy(String, String...) - Method in class org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so that we can run aggregation on them.
- groupBy(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so we can run aggregation on them.
- groupBy(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so that we can run aggregation on them.
- groupByExpressions() - Method in class org.apache.spark.sql.connector.expressions.aggregate.Aggregation
-
- groupByKey(Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey() - Method in class org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(Partitioner) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(int) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey() - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(Function1<T, K>, Encoder<K>) - Method in class org.apache.spark.sql.Dataset
-
- groupByKey(MapFunction<T, K>, Encoder<K>) - Method in class org.apache.spark.sql.Dataset
-
- groupByKey() - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
to each RDD.
- groupByKey(int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
to each RDD.
- groupByKey(Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
on each RDD of this
DStream.
- groupByKey() - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
to each RDD.
- groupByKey(int) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
to each RDD.
- groupByKey(Partitioner) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
on each RDD.
- groupByKeyAndWindow(Duration) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
over a sliding window.
- groupByKeyAndWindow(Duration, Duration) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
over a sliding window.
- groupByKeyAndWindow(Duration, Duration, int) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
over a sliding window on this
DStream.
- groupByKeyAndWindow(Duration, Duration, Partitioner) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey
over a sliding window on this
DStream.
- groupByKeyAndWindow(Duration) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
over a sliding window.
- groupByKeyAndWindow(Duration, Duration) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
over a sliding window.
- groupByKeyAndWindow(Duration, Duration, int) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey
over a sliding window on this
DStream.
- groupByKeyAndWindow(Duration, Duration, Partitioner) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Create a new DStream by applying groupByKey
over a sliding window on this
DStream.
- groupByPositionRangeError(int, int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- groupByPositionRefersToAggregateFunctionError(int, Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- GroupByType$() - Constructor for class org.apache.spark.sql.RelationalGroupedDataset.GroupByType$
-
- groupEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.Graph
-
Merges multiple edges between two vertices into a single edge.
- groupEdges(Function2<ED, ED, ED>) - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- groupHash() - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
- grouping(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated
or not, returns 1 for aggregated or 0 for not aggregated in the result set.
- grouping(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated
or not, returns 1 for aggregated or 0 for not aggregated in the result set.
- grouping_id(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the level of grouping, equals to
- grouping_id(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the level of grouping, equals to
- groupingColInvalidError(Expression, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- groupingIDMismatchError(GroupingID, Seq<Expression>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- groupingMustWithGroupingSetsOrCubeOrRollupError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- groupingSizeTooLargeError(int) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- GroupMappingServiceProvider - Interface in org.apache.spark.security
-
This Spark trait is used for mapping a given userName to a set of groups which it belongs to.
- GroupState<S> - Interface in org.apache.spark.sql.streaming
-
:: Experimental ::
- GroupStateTimeout - Class in org.apache.spark.sql.streaming
-
Represents the type of timeouts possible for the Dataset operations
mapGroupsWithState
and flatMapGroupsWithState
.
- GroupStateTimeout() - Constructor for class org.apache.spark.sql.streaming.GroupStateTimeout
-
- groupWith(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- gt(double) - Static method in class org.apache.spark.ml.param.ParamValidators
-
Check if value is greater than lowerBound
- gt(Object) - Method in class org.apache.spark.sql.Column
-
Greater than.
- gt(T, T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
-
- gt(T, T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
-
- gtEq(double) - Static method in class org.apache.spark.ml.param.ParamValidators
-
Check if value is greater than or equal to lowerBound
- gteq(T, T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
-
- gteq(T, T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
-
- guard(Function0<Parsers.Parser<T>>) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- s() - Method in class org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- safeCall(Function0<T>) - Method in interface org.apache.spark.security.CryptoStreamUtils.BaseErrorHandler
-
- safeDoubleToJValue(double) - Static method in class org.apache.spark.sql.streaming.SafeJsonSerializer
-
- SafeJsonSerializer - Class in org.apache.spark.sql.streaming
-
- SafeJsonSerializer() - Constructor for class org.apache.spark.sql.streaming.SafeJsonSerializer
-
- safeMapToJValue(Map<String, T>, Function1<T, JsonAST.JValue>) - Static method in class org.apache.spark.sql.streaming.SafeJsonSerializer
-
Convert map to JValue while handling empty maps.
- sameSemantics(Dataset<T>) - Method in class org.apache.spark.sql.Dataset
-
Returns
true
when the logical query plans inside both
Dataset
s are equal and
therefore return same results.
- sameThread() - Static method in class org.apache.spark.util.ThreadUtils
-
An ExecutionContextExecutor
that runs each task in the thread that invokes execute/submit
.
- sameThreadExecutorService() - Static method in class org.apache.spark.util.ThreadUtils
-
- sample(boolean, Double) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return a sampled subset of this RDD.
- sample(boolean, Double, long) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double, long) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double) - Method in class org.apache.spark.api.java.JavaRDD
-
Return a sampled subset of this RDD with a random seed.
- sample(boolean, double, long) - Method in class org.apache.spark.api.java.JavaRDD
-
Return a sampled subset of this RDD, with a user-supplied seed.
- sample(boolean, double, long) - Method in class org.apache.spark.rdd.RDD
-
Return a sampled subset of this RDD.
- sample(double, long) - Method in class org.apache.spark.sql.Dataset
-
Returns a new
Dataset
by sampling a fraction of rows (without replacement),
using a user-supplied seed.
- sample(double) - Method in class org.apache.spark.sql.Dataset
-
Returns a new
Dataset
by sampling a fraction of rows (without replacement),
using a random seed.
- sample(boolean, double, long) - Method in class org.apache.spark.sql.Dataset
-
Returns a new
Dataset
by sampling a fraction of rows, using a user-supplied seed.
- sample(boolean, double) - Method in class org.apache.spark.sql.Dataset
-
Returns a new
Dataset
by sampling a fraction of rows, using a random seed.
- sample() - Method in class org.apache.spark.util.random.BernoulliCellSampler
-
- sample() - Method in class org.apache.spark.util.random.BernoulliSampler
-
- sample() - Method in class org.apache.spark.util.random.PoissonSampler
-
- sample(Iterator<T>) - Method in class org.apache.spark.util.random.PoissonSampler
-
- sample(Iterator<T>) - Method in interface org.apache.spark.util.random.RandomSampler
-
take a random sample
- sample() - Method in interface org.apache.spark.util.random.RandomSampler
-
Whether to sample the next item or not.
- sampleBy(String, Map<T, Object>, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Returns a stratified sample without replacement based on the fraction given on each stratum.
- sampleBy(String, Map<T, Double>, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Returns a stratified sample without replacement based on the fraction given on each stratum.
- sampleBy(Column, Map<T, Object>, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
Returns a stratified sample without replacement based on the fraction given on each stratum.
- sampleBy(Column, Map<T, Double>, long) - Method in class org.apache.spark.sql.DataFrameStatFunctions
-
(Java-specific) Returns a stratified sample without replacement based on the fraction given
on each stratum.
- sampleByKey(boolean, Map<K, Double>, long) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKey(boolean, Map<K, Double>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKey(boolean, Map<K, Object>, long) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKeyExact(boolean, Map<K, Double>, long) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- sampleByKeyExact(boolean, Map<K, Double>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- sampleByKeyExact(boolean, Map<K, Object>, long) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- SamplePathFilter - Class in org.apache.spark.ml.image
-
Filter that allows loading a fraction of HDFS files.
- SamplePathFilter() - Constructor for class org.apache.spark.ml.image.SamplePathFilter
-
- samplePointsPerPartitionHint() - Method in class org.apache.spark.RangePartitioner
-
- sampleRatio() - Method in class org.apache.spark.ml.image.SamplePathFilter
-
- sampleStdev() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Compute the sample standard deviation of this RDD's elements (which corrects for bias in
estimating the standard deviation by dividing by N-1 instead of N).
- sampleStdev() - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the sample standard deviation of this RDD's elements (which corrects for bias in
estimating the standard deviation by dividing by N-1 instead of N).
- sampleStdev() - Method in class org.apache.spark.util.StatCounter
-
Return the sample standard deviation of the values, which corrects for bias in estimating the
variance by dividing by N-1 instead of N.
- sampleVariance() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Compute the sample variance of this RDD's elements (which corrects for bias in
estimating the standard variance by dividing by N-1 instead of N).
- sampleVariance() - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the sample variance of this RDD's elements (which corrects for bias in
estimating the variance by dividing by N-1 instead of N).
- sampleVariance() - Method in class org.apache.spark.util.StatCounter
-
Return the sample variance, which corrects for bias in estimating the variance by dividing
by N-1 instead of N.
- SamplingUtils - Class in org.apache.spark.util.random
-
- SamplingUtils() - Constructor for class org.apache.spark.util.random.SamplingUtils
-
- sanitizeDirName(String) - Static method in class org.apache.spark.util.Utils
-
- save(String) - Method in class org.apache.spark.ml.feature.HashingTF
-
- save(String) - Method in interface org.apache.spark.ml.util.MLWritable
-
Saves this ML instance to the input path, a shortcut of write.save(path)
.
- save(String) - Method in class org.apache.spark.ml.util.MLWriter
-
Saves the ML instances to the input path.
- save(SparkContext, String, String, int, int, Vector, double, Option<Object>) - Method in class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
Helper method for saving GLM classification model metadata and data.
- save(SparkContext, String) - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.classification.NaiveBayesModel
-
- save(SparkContext, String, org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0.Data) - Method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- save(SparkContext, String, org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0.Data) - Method in class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.classification.SVMModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel
-
- save(SparkContext, BisectingKMeansModel, String) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
- save(SparkContext, BisectingKMeansModel, String) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV2_0$
-
- save(SparkContext, BisectingKMeansModel, String) - Method in class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV3_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.DistributedLDAModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.KMeansModel
-
- save(SparkContext, KMeansModel, String) - Method in class org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
- save(SparkContext, KMeansModel, String) - Method in class org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV2_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- save(SparkContext, PowerIterationClusteringModel, String) - Method in class org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- save(SparkContext, ChiSqSelectorModel, String) - Method in class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.feature.Word2VecModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.fpm.FPGrowthModel
-
Save this model to the given path.
- save(FPGrowthModel<?>, String) - Method in class org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.fpm.PrefixSpanModel
-
Save this model to the given path.
- save(PrefixSpanModel<?>, String) - Method in class org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Save this model to the given path.
- save(MatrixFactorizationModel, String) - Method in class org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
Saves a
MatrixFactorizationModel
, where user features are saved under
data/users
and
product features are saved under
data/products
.
- save(SparkContext, String, String, Vector, double) - Method in class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
Helper method for saving GLM regression model metadata and data.
- save(SparkContext, String) - Method in class org.apache.spark.mllib.regression.IsotonicRegressionModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.regression.LassoModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.regression.LinearRegressionModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.regression.RidgeRegressionModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- save(SparkContext, String, DecisionTreeModel) - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- save(SparkContext, String) - Method in class org.apache.spark.mllib.tree.model.RandomForestModel
-
- save(SparkContext, String) - Method in interface org.apache.spark.mllib.util.Saveable
-
Save this model to the given path.
- save(String) - Method in class org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame
at the specified path.
- save() - Method in class org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame
as the specified table.
- Saveable - Interface in org.apache.spark.mllib.util
-
Trait for models and transformers which may be saved as files.
- saveAsHadoopDataset(JobConf) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported storage system, using a Hadoop JobConf object for
that storage system.
- saveAsHadoopDataset(JobConf) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported storage system, using a Hadoop JobConf object for
that storage system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>, JobConf) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>, Class<? extends CompressionCodec>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system, compressing with the supplied codec.
- saveAsHadoopFile(String, ClassTag<F>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat
class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<? extends CompressionCodec>, ClassTag<F>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat
class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Class<? extends CompressionCodec>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat
class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, JobConf, Option<Class<? extends CompressionCodec>>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat
class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFiles(String, String) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, JobConf) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, ClassTag<F>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, JobConf) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsLibSVMFile(RDD<LabeledPoint>, String) - Static method in class org.apache.spark.mllib.util.MLUtils
-
Save labeled data in LIBSVM format.
- saveAsNewAPIHadoopDataset(Configuration) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported storage system, using
a Configuration object for that storage system.
- saveAsNewAPIHadoopDataset(Configuration) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported storage system with new Hadoop API, using a Hadoop
Configuration object for that storage system.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<F>, Configuration) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<F>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsNewAPIHadoopFile(String, ClassTag<F>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a new Hadoop API OutputFormat
(mapreduce.OutputFormat) object supporting the key and value types K and V in this RDD.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Configuration) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a new Hadoop API OutputFormat
(mapreduce.OutputFormat) object supporting the key and value types K and V in this RDD.
- saveAsNewAPIHadoopFiles(String, String) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, Configuration) - Method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, ClassTag<F>) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Configuration) - Method in class org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this
DStream as a Hadoop file.
- saveAsObjectFile(String) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a SequenceFile of serialized objects.
- saveAsObjectFile(String) - Method in class org.apache.spark.rdd.RDD
-
Save this RDD as a SequenceFile of serialized objects.
- saveAsObjectFiles(String, String) - Method in class org.apache.spark.streaming.dstream.DStream
-
Save each RDD in this DStream as a Sequence file of serialized objects.
- saveAsSequenceFile(String, Option<Class<? extends CompressionCodec>>) - Method in class org.apache.spark.rdd.SequenceFileRDDFunctions
-
Output the RDD as a Hadoop SequenceFile using the Writable types we infer from the RDD's key
and value types.
- saveAsTable(String) - Method in class org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame
as the specified table.
- saveAsTextFile(String) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a text file, using string representations of elements.
- saveAsTextFile(String, Class<? extends CompressionCodec>) - Method in interface org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a compressed text file, using string representations of elements.
- saveAsTextFile(String) - Method in class org.apache.spark.rdd.RDD
-
Save this RDD as a text file, using string representations of elements.
- saveAsTextFile(String, Class<? extends CompressionCodec>) - Method in class org.apache.spark.rdd.RDD
-
Save this RDD as a compressed text file, using string representations of elements.
- saveAsTextFiles(String, String) - Method in class org.apache.spark.streaming.dstream.DStream
-
Save each RDD in this DStream as at text file, using string representation
of elements.
- saveDataIntoViewNotAllowedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- savedTasks() - Method in class org.apache.spark.status.LiveStage
-
- saveImpl(Params, PipelineStage[], SparkContext, String) - Method in class org.apache.spark.ml.Pipeline.SharedReadWrite$
-
Save metadata and stages for a
Pipeline
or
PipelineModel
- save metadata to path/metadata
- save stages to stages/IDX_UID
- saveImpl(M, String, SparkSession, JsonAST.JObject) - Static method in class org.apache.spark.ml.tree.EnsembleModelReadWrite
-
Helper method for saving a tree ensemble to disk.
- SaveInstanceEnd - Class in org.apache.spark.ml
-
Event fired after MLWriter.save
.
- SaveInstanceEnd(String) - Constructor for class org.apache.spark.ml.SaveInstanceEnd
-
- SaveInstanceStart - Class in org.apache.spark.ml
-
Event fired before MLWriter.save
.
- SaveInstanceStart(String) - Constructor for class org.apache.spark.ml.SaveInstanceStart
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - Constructor for class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- SaveLoadV2_0$() - Constructor for class org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- SaveLoadV2_0$() - Constructor for class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV2_0$
-
- SaveLoadV2_0$() - Constructor for class org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV2_0$
-
- SaveLoadV3_0$() - Constructor for class org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV3_0$
-
- SaveMode - Enum in org.apache.spark.sql
-
SaveMode is used to specify the expected behavior of saving a DataFrame to a data source.
- saveModeUnsupportedError(Object, boolean) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- sc() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- sc() - Method in interface org.apache.spark.ml.util.BaseReadWrite
-
Returns the underlying `SparkContext`.
- sc() - Method in class org.apache.spark.sql.SQLImplicits.StringToColumn
-
- scal(double, Vector) - Static method in class org.apache.spark.ml.linalg.BLAS
-
x = a * x
- scal(double, Vector) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
x = a * x
- SCALA_VERSION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo
-
- scalaBoolean() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive boolean type.
- scalaByte() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive byte type.
- scalaDouble() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive double type.
- scalaFloat() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive float type.
- scalaInt() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive int type.
- scalaIntToJavaLong(DStream<Object>) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
- scalaLong() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive long type.
- ScalarFunction<R> - Interface in org.apache.spark.sql.connector.catalog.functions
-
Interface for a function that produces a result value for each input row.
- scalaShort() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive short type.
- scalaToJavaLong(JavaPairDStream<K, Object>, ClassTag<K>) - Static method in class org.apache.spark.streaming.api.java.JavaPairDStream
-
- scalaVersion() - Method in class org.apache.spark.status.api.v1.RuntimeInfo
-
- scale() - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- scale() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- scale() - Method in class org.apache.spark.mllib.random.GammaGenerator
-
- scale() - Method in class org.apache.spark.sql.types.Decimal
-
- scale() - Method in class org.apache.spark.sql.types.DecimalType
-
- scalingVec() - Method in class org.apache.spark.ml.feature.ElementwiseProduct
-
the vector to multiply with input vectors
- scalingVec() - Method in class org.apache.spark.mllib.feature.ElementwiseProduct
-
- Scan - Interface in org.apache.spark.sql.connector.read
-
A logical representation of a data source scan.
- Scan.ColumnarSupportMode - Enum in org.apache.spark.sql.connector.read
-
This enum defines how the columnar support for the partitions of the data source
should be determined.
- ScanBuilder - Interface in org.apache.spark.sql.connector.read
-
An interface for building the
Scan
.
- Schedulable - Interface in org.apache.spark.scheduler
-
An interface for schedulable entities.
- SchedulableBuilder - Interface in org.apache.spark.scheduler
-
An interface to build Schedulable tree
buildPools: build the tree nodes(pools)
addTaskSetManager: build the leaf nodes(TaskSetManagers)
- schedulableQueue() - Method in interface org.apache.spark.scheduler.Schedulable
-
- SCHEDULED() - Static method in class org.apache.spark.streaming.scheduler.ReceiverState
-
- SCHEDULER_DELAY() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SCHEDULER_DELAY() - Static method in class org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SCHEDULER_DELAY() - Static method in class org.apache.spark.ui.ToolTips
-
- SCHEDULER_DELAY_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SCHEDULER_DELAY_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- SCHEDULER_DELAY_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- SchedulerBackend - Interface in org.apache.spark.scheduler
-
A backend interface for scheduling systems that allows plugging in different ones under
TaskSchedulerImpl.
- SchedulerBackendUtils - Class in org.apache.spark.scheduler.cluster
-
- SchedulerBackendUtils() - Constructor for class org.apache.spark.scheduler.cluster.SchedulerBackendUtils
-
- schedulerDelay() - Method in class org.apache.spark.status.api.v1.TaskData
-
- schedulerDelay() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
-
- schedulerDelay(TaskData) - Static method in class org.apache.spark.status.AppStatusUtils
-
- schedulerDelay(long, long, long, long, long, long) - Static method in class org.apache.spark.status.AppStatusUtils
-
- SchedulerPool - Class in org.apache.spark.status
-
- SchedulerPool(String) - Constructor for class org.apache.spark.status.SchedulerPool
-
- SCHEDULING_POOL_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SchedulingAlgorithm - Interface in org.apache.spark.scheduler
-
An interface for sort algorithm
FIFO: FIFO algorithm between TaskSetManagers
FS: FS algorithm between Pools, and FIFO or FS within Pools
- schedulingDelay() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
-
- schedulingDelay() - Method in class org.apache.spark.streaming.scheduler.BatchInfo
-
Time taken for the first job of this batch to start processing from the time this batch
was submitted to the streaming scheduler.
- schedulingMode() - Method in interface org.apache.spark.scheduler.Schedulable
-
- SchedulingMode - Class in org.apache.spark.scheduler
-
"FAIR" and "FIFO" determines which policy is used
to order tasks amongst a Schedulable's sub-queues
"NONE" is used when the a Schedulable has no sub-queues.
- SchedulingMode() - Constructor for class org.apache.spark.scheduler.SchedulingMode
-
- schedulingMode() - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- schedulingPool() - Method in class org.apache.spark.status.api.v1.StageData
-
- schedulingPool() - Method in class org.apache.spark.status.LiveStage
-
- schema() - Method in interface org.apache.spark.sql.connector.catalog.Table
-
Deprecated.
- schema() - Method in interface org.apache.spark.sql.connector.catalog.View
-
The schema for the view when the view is created after applying column aliases.
- schema() - Method in interface org.apache.spark.sql.connector.write.LogicalWriteInfo
-
the schema of the input data from Spark to data source.
- schema(StructType) - Method in class org.apache.spark.sql.DataFrameReader
-
Specifies the input schema.
- schema(String) - Method in class org.apache.spark.sql.DataFrameReader
-
Specifies the schema by using the input DDL-formatted string.
- schema() - Method in class org.apache.spark.sql.Dataset
-
Returns the schema of this Dataset.
- schema() - Method in interface org.apache.spark.sql.Encoder
-
Returns the schema of encoding this type of object as a Row.
- schema() - Method in interface org.apache.spark.sql.Row
-
Schema for the row.
- schema() - Method in class org.apache.spark.sql.sources.BaseRelation
-
- schema(StructType) - Method in class org.apache.spark.sql.streaming.DataStreamReader
-
Specifies the input schema.
- schema(String) - Method in class org.apache.spark.sql.streaming.DataStreamReader
-
Specifies the schema by using the input DDL-formatted string.
- schema_of_csv(String) - Static method in class org.apache.spark.sql.functions
-
Parses a CSV string and infers its schema in DDL format.
- schema_of_csv(Column) - Static method in class org.apache.spark.sql.functions
-
Parses a CSV string and infers its schema in DDL format.
- schema_of_csv(Column, Map<String, String>) - Static method in class org.apache.spark.sql.functions
-
Parses a CSV string and infers its schema in DDL format using options.
- schema_of_json(String) - Static method in class org.apache.spark.sql.functions
-
Parses a JSON string and infers its schema in DDL format.
- schema_of_json(Column) - Static method in class org.apache.spark.sql.functions
-
Parses a JSON string and infers its schema in DDL format.
- schema_of_json(Column, Map<String, String>) - Static method in class org.apache.spark.sql.functions
-
Parses a JSON string and infers its schema in DDL format using options.
- SchemaConverters - Class in org.apache.spark.sql.avro
-
This object contains method that are used to convert sparkSQL schemas to avro schemas and vice
versa.
- SchemaConverters() - Constructor for class org.apache.spark.sql.avro.SchemaConverters
-
- SchemaConverters.SchemaType - Class in org.apache.spark.sql.avro
-
Internal wrapper for SQL data type and nullability.
- SchemaConverters.SchemaType$ - Class in org.apache.spark.sql.avro
-
- schemaFailToParseError(String, Throwable) - Static method in class org.apache.spark.sql.errors.DataTypeErrors
-
- schemaIsNotStructTypeError(Expression, DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- schemaNotSpecifiedForSchemaRelationProviderError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- SchemaRelationProvider - Interface in org.apache.spark.sql.sources
-
Implemented by objects that produce relations for a specific kind of data source
with a given schema.
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- schemasExists(Connection, JDBCOptions, String) - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Check schema exists or not.
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- schemasExists(Connection, JDBCOptions, String) - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- SchemaType(DataType, boolean) - Constructor for class org.apache.spark.sql.avro.SchemaConverters.SchemaType
-
- SchemaType$() - Constructor for class org.apache.spark.sql.avro.SchemaConverters.SchemaType$
-
- SchemaUtils - Class in org.apache.spark.ml.util
-
Utils for handling schemas.
- SchemaUtils() - Constructor for class org.apache.spark.ml.util.SchemaUtils
-
- SchemaUtils - Class in org.apache.spark.sql.util
-
Utils for handling schemas.
- SchemaUtils() - Constructor for class org.apache.spark.sql.util.SchemaUtils
-
- scope() - Method in class org.apache.spark.storage.RDDInfo
-
- scoreAndLabels() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
- scoreCol() - Method in interface org.apache.spark.ml.classification.BinaryClassificationSummary
-
Field in "predictions" which gives the probability or rawPrediction of each class as a
vector.
- scoreCol() - Method in interface org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
- scoreCol() - Method in class org.apache.spark.ml.classification.BinaryRandomForestClassificationSummaryImpl
-
- scoreCol() - Method in class org.apache.spark.ml.classification.FMClassificationSummaryImpl
-
- scoreCol() - Method in class org.apache.spark.ml.classification.LinearSVCSummaryImpl
-
- scoreLabelsWeight() - Method in class org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
- scratch() - Method in class org.apache.spark.mllib.optimization.NNLS.Workspace
-
- sec(Column) - Static method in class org.apache.spark.sql.functions
-
- second(Column) - Static method in class org.apache.spark.sql.functions
-
Extracts the seconds as an integer from a given date/timestamp/string.
- SECOND() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- secondArgumentInFunctionIsNotBooleanLiteralError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- secondArgumentNotDoubleLiteralError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- secondArgumentOfFunctionIsNotIntegerError(String, NumberFormatException) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- seconds() - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- seconds(long) - Static method in class org.apache.spark.streaming.Durations
-
- Seconds - Class in org.apache.spark.streaming
-
Helper object that creates instance of
Duration
representing
a given number of seconds.
- Seconds() - Constructor for class org.apache.spark.streaming.Seconds
-
- SecurityConfigurationLock - Class in org.apache.spark.security
-
There are cases when global JVM security configuration must be modified.
- SecurityConfigurationLock() - Constructor for class org.apache.spark.security.SecurityConfigurationLock
-
- securityManager() - Method in class org.apache.spark.SparkEnv
-
- securityManager() - Method in interface org.apache.spark.status.api.v1.UIRoot
-
- SecurityUtils - Class in org.apache.spark.util
-
Various utility methods used by Spark Security.
- SecurityUtils() - Constructor for class org.apache.spark.util.SecurityUtils
-
- seed() - Method in class org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- seed() - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- seed() - Method in class org.apache.spark.ml.classification.FMClassificationModel
-
- seed() - Method in class org.apache.spark.ml.classification.FMClassifier
-
- seed() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- seed() - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- seed() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- seed() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- seed() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
- seed() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- seed() - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- seed() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- seed() - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- seed() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- seed() - Method in class org.apache.spark.ml.clustering.KMeans
-
- seed() - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- seed() - Method in class org.apache.spark.ml.clustering.LDA
-
- seed() - Method in class org.apache.spark.ml.clustering.LDAModel
-
- seed() - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- seed() - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- seed() - Method in class org.apache.spark.ml.feature.Word2Vec
-
- seed() - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- seed() - Method in interface org.apache.spark.ml.param.shared.HasSeed
-
Param for random seed.
- seed() - Method in class org.apache.spark.ml.recommendation.ALS
-
- seed() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- seed() - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- seed() - Method in class org.apache.spark.ml.regression.FMRegressionModel
-
- seed() - Method in class org.apache.spark.ml.regression.FMRegressor
-
- seed() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- seed() - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- seed() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
-
- seed() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- seed() - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- seed() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
-
- seed() - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- seed() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- seedParam() - Static method in class org.apache.spark.ml.image.SamplePathFilter
-
- select(Column...) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of column based expressions.
- select(String, String...) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of columns.
- select(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of column based expressions.
- select(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of columns.
- select(TypedColumn<T, U1>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset by computing the given
Column
expression for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset by computing the given
Column
expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset by computing the given
Column
expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>, TypedColumn<T, U4>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset by computing the given
Column
expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>, TypedColumn<T, U4>, TypedColumn<T, U5>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset by computing the given
Column
expressions for each element.
- selectedFeatures() - Method in class org.apache.spark.ml.feature.ChiSqSelectorModel
-
- selectedFeatures() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- selectedFeatures() - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
-
- selectedFeatures() - Method in class org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- selectExpr(String...) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of SQL expressions.
- selectExpr(Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Selects a set of SQL expressions.
- selectExprNotInGroupByError(Expression, Seq<Alias>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- selectionMode() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- selectionMode() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- selectionMode() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
The selection mode.
- selectionThreshold() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- selectionThreshold() - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- selectionThreshold() - Method in interface org.apache.spark.ml.feature.UnivariateFeatureSelectorParams
-
The upper bound of the features that selector will select.
- SelectorParams - Interface in org.apache.spark.ml.feature
-
Params for Selector
and SelectorModel
.
- selectorType() - Method in interface org.apache.spark.ml.feature.SelectorParams
-
The selector type.
- selectorType() - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- semanticHash() - Method in class org.apache.spark.sql.Dataset
-
Returns a
hashCode
of the logical query plan against this
Dataset
.
- send(Object) - Method in interface org.apache.spark.api.plugin.PluginContext
-
Send a message to the plugin's driver-side component.
- sendData(String, Seq<Object>) - Method in interface org.apache.spark.streaming.kinesis.KinesisDataGenerator
-
Sends the data to Kinesis and returns the metadata for everything that has been sent.
- sender() - Method in class org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
-
- sendFetchMergedStatusRequest(ShuffleBlockFetcherIterator.FetchRequest) - Method in class org.apache.spark.storage.PushBasedFetchHelper
-
This is executed by the task thread when the iterator is initialized and only if it has
push-merged blocks for which it needs to fetch the metadata.
- sendResubmittedTaskStatusForShuffleMapStagesOnlyError() - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- sendToDst(A) - Method in class org.apache.spark.graphx.EdgeContext
-
Sends a message to the destination vertex.
- sendToDst(A) - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- sendToSrc(A) - Method in class org.apache.spark.graphx.EdgeContext
-
Sends a message to the source vertex.
- sendToSrc(A) - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- sentences(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Splits a string into arrays of sentences, where each sentence is an array of words.
- sentences(Column) - Static method in class org.apache.spark.sql.functions
-
Splits a string into arrays of sentences, where each sentence is an array of words.
- sequence() - Method in class org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
-
- sequence(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Generate a sequence of integers from start to stop, incrementing by step.
- sequence(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Generate a sequence of integers from start to stop,
incrementing by 1 if start is less than or equal to stop, otherwise -1.
- sequenceCol() - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
Param for the name of the sequence column in dataset (default "sequence"), rows with
nulls in this column are ignored.
- sequenceFile(String, Class<K>, Class<V>, int) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, Class<K>, Class<V>) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Get an RDD for a Hadoop SequenceFile.
- sequenceFile(String, Class<K>, Class<V>, int) - Method in class org.apache.spark.SparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, Class<K>, Class<V>) - Method in class org.apache.spark.SparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, int, ClassTag<K>, ClassTag<V>, Function0<WritableConverter<K>>, Function0<WritableConverter<V>>) - Method in class org.apache.spark.SparkContext
-
Version of sequenceFile() for types implicitly convertible to Writables through a
WritableConverter.
- SequenceFileRDDFunctions<K,V> - Class in org.apache.spark.rdd
-
Extra functions available on RDDs of (key, value) pairs to create a Hadoop SequenceFile,
through an implicit conversion.
- SequenceFileRDDFunctions(RDD<Tuple2<K, V>>, Class<? extends Writable>, Class<? extends Writable>, Function1<K, Writable>, ClassTag<K>, Function1<V, Writable>, ClassTag<V>) - Constructor for class org.apache.spark.rdd.SequenceFileRDDFunctions
-
- SER_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SerDe - Class in org.apache.spark.api.r
-
Utility functions to serialize, deserialize objects to / from R
- SerDe() - Constructor for class org.apache.spark.api.r.SerDe
-
- serDeInterfaceNotFoundError(NoClassDefFoundError) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- SerializableConfiguration - Class in org.apache.spark.util
-
Hadoop configuration but serializable.
- SerializableConfiguration(Configuration) - Constructor for class org.apache.spark.util.SerializableConfiguration
-
- SerializableMapWrapper(Map<A, B>) - Constructor for class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- SerializableWritable<T extends org.apache.hadoop.io.Writable> - Class in org.apache.spark
-
- SerializableWritable(T) - Constructor for class org.apache.spark.SerializableWritable
-
- SerializationDebugger - Class in org.apache.spark.serializer
-
- SerializationDebugger() - Constructor for class org.apache.spark.serializer.SerializationDebugger
-
- SerializationDebugger.ObjectStreamClassMethods - Class in org.apache.spark.serializer
-
An implicit class that allows us to call private methods of ObjectStreamClass.
- SerializationDebugger.ObjectStreamClassMethods$ - Class in org.apache.spark.serializer
-
- SerializationFormats - Class in org.apache.spark.api.r
-
- SerializationFormats() - Constructor for class org.apache.spark.api.r.SerializationFormats
-
- SerializationStream - Class in org.apache.spark.serializer
-
:: DeveloperApi ::
A stream for writing serialized objects.
- SerializationStream() - Constructor for class org.apache.spark.serializer.SerializationStream
-
- serializationStream() - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
-
- serialize(Vector) - Method in class org.apache.spark.mllib.linalg.VectorUDT
-
- serialize(T, ClassTag<T>) - Method in class org.apache.spark.serializer.DummySerializerInstance
-
- serialize(T, ClassTag<T>) - Method in class org.apache.spark.serializer.SerializerInstance
-
- serialize(UserType) - Method in class org.apache.spark.sql.types.UserDefinedType
-
Convert the user type to a SQL datum
- serialize(AccumulableInfo) - Static method in class org.apache.spark.status.protobuf.AccumulableInfoSerializer
-
- serialize(Enumeration.Value) - Static method in class org.apache.spark.status.protobuf.DeterministicLevelSerializer
-
- serialize(ExecutorMetrics) - Static method in class org.apache.spark.status.protobuf.ExecutorMetricsSerializer
-
- serialize(ExecutorStageSummary) - Static method in class org.apache.spark.status.protobuf.ExecutorStageSummarySerializer
-
- serialize(JobExecutionStatus) - Static method in class org.apache.spark.status.protobuf.JobExecutionStatusSerializer
-
- serialize(T) - Method in interface org.apache.spark.status.protobuf.ProtobufSerDe
-
Serialize the input data of the type T
to Array[Byte]
.
- serialize(SinkProgress) - Static method in class org.apache.spark.status.protobuf.sql.SinkProgressSerializer
-
- serialize(SourceProgress) - Static method in class org.apache.spark.status.protobuf.sql.SourceProgressSerializer
-
- serialize(SQLPlanMetric) - Static method in class org.apache.spark.status.protobuf.sql.SQLPlanMetricSerializer
-
- serialize(StateOperatorProgress) - Static method in class org.apache.spark.status.protobuf.sql.StateOperatorProgressSerializer
-
- serialize(StreamingQueryProgress) - Static method in class org.apache.spark.status.protobuf.sql.StreamingQueryProgressSerializer
-
- serialize(StageStatus) - Static method in class org.apache.spark.status.protobuf.StageStatusSerializer
-
- serialize(T) - Method in interface org.apache.spark.util.SparkSerDeUtils
-
Serialize an object using Java serialization
- serialize(T) - Static method in class org.apache.spark.util.Utils
-
- SERIALIZED_R_DATA_SCHEMA() - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- serializedData() - Method in class org.apache.spark.scheduler.local.StatusUpdate
-
- serializedMapAndMergeStatus(org.apache.spark.broadcast.BroadcastManager, boolean, int, SparkConf) - Method in class org.apache.spark.ShuffleStatus
-
Serializes the mapStatuses and mergeStatuses array into an efficient compressed format.
- serializedMapStatus(org.apache.spark.broadcast.BroadcastManager, boolean, int, SparkConf) - Method in class org.apache.spark.ShuffleStatus
-
Serializes the mapStatuses array into an efficient compressed format.
- SerializedMemoryEntry<T> - Class in org.apache.spark.storage.memory
-
- SerializedMemoryEntry(ChunkedByteBuffer, MemoryMode, ClassTag<T>) - Constructor for class org.apache.spark.storage.memory.SerializedMemoryEntry
-
- serializedPyClass() - Method in class org.apache.spark.sql.types.UserDefinedType
-
Serialized Python UDT class, if exists.
- SerializedValuesHolder<T> - Class in org.apache.spark.storage.memory
-
A holder for storing the serialized values.
- SerializedValuesHolder(BlockId, int, ClassTag<T>, MemoryMode, org.apache.spark.serializer.SerializerManager) - Constructor for class org.apache.spark.storage.memory.SerializedValuesHolder
-
- Serializer - Class in org.apache.spark.serializer
-
:: DeveloperApi ::
A serializer.
- Serializer() - Constructor for class org.apache.spark.serializer.Serializer
-
- serializer() - Method in class org.apache.spark.ShuffleDependency
-
- serializer() - Method in class org.apache.spark.SparkEnv
-
- serializerForHistoryServer(SparkConf) - Static method in class org.apache.spark.status.KVUtils
-
- SerializerHelper - Class in org.apache.spark.serializer
-
- SerializerHelper() - Constructor for class org.apache.spark.serializer.SerializerHelper
-
- SerializerInstance - Class in org.apache.spark.serializer
-
:: DeveloperApi ::
An instance of a serializer, for use by one thread at a time.
- SerializerInstance() - Constructor for class org.apache.spark.serializer.SerializerInstance
-
- serializerManager() - Method in class org.apache.spark.SparkEnv
-
- serializeStream(OutputStream) - Method in class org.apache.spark.serializer.DummySerializerInstance
-
- serializeStream(OutputStream) - Method in class org.apache.spark.serializer.SerializerInstance
-
- serializeToChunkedBuffer(SerializerInstance, T, long, ClassTag<T>) - Static method in class org.apache.spark.serializer.SerializerHelper
-
- serializeViaNestedStream(OutputStream, SerializerInstance, Function1<SerializationStream, BoxedUnit>) - Static method in class org.apache.spark.util.Utils
-
Serialize via nested stream using specific serializer
- serviceName() - Method in interface org.apache.spark.security.HadoopDelegationTokenProvider
-
Name of the service to provide delegation tokens.
- servletContext() - Method in interface org.apache.spark.status.api.v1.ApiRequestContext
-
- ServletParams(Function1<HttpServletRequest, T>, String, Function1<T, String>) - Constructor for class org.apache.spark.ui.JettyUtils.ServletParams
-
- ServletParams$() - Constructor for class org.apache.spark.ui.JettyUtils.ServletParams$
-
- session(SparkSession) - Static method in class org.apache.spark.ml.r.RWrappers
-
- session(SparkSession) - Method in interface org.apache.spark.ml.util.BaseReadWrite
-
Sets the Spark Session to use for saving/loading.
- session(SparkSession) - Method in class org.apache.spark.ml.util.GeneralMLWriter
-
- session(SparkSession) - Method in class org.apache.spark.ml.util.MLReader
-
- session(SparkSession) - Method in class org.apache.spark.ml.util.MLWriter
-
- session_window(Column, String) - Static method in class org.apache.spark.sql.functions
-
Generates session window given a timestamp specifying column.
- session_window(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Generates session window given a timestamp specifying column.
- SessionCatalogAndIdentifier() - Method in interface org.apache.spark.sql.connector.catalog.LookupCatalog
-
- SessionCatalogAndIdentifier() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier
-
- SessionCatalogAndIdentifier$() - Constructor for class org.apache.spark.sql.connector.catalog.LookupCatalog.SessionCatalogAndIdentifier$
-
- SessionConfigSupport - Interface in org.apache.spark.sql.connector.catalog
-
- sessionId() - Method in class org.apache.spark.storage.CacheId
-
- sessionState() - Method in class org.apache.spark.sql.SparkSession
-
- sessionWindowGapDurationDataTypeError(DataType) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- set(long, long, int, int, VD, VD, ED) - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- Set() - Static method in class org.apache.spark.metrics.sink.StatsdMetricType
-
- set(Param<T>, T) - Method in interface org.apache.spark.ml.param.Params
-
Sets a parameter in the embedded param map.
- set(String, Object) - Method in interface org.apache.spark.ml.param.Params
-
Sets a parameter (by name) in the embedded param map.
- set(ParamPair<?>) - Method in interface org.apache.spark.ml.param.Params
-
Sets a parameter in the embedded param map.
- set(String, long, long) - Static method in class org.apache.spark.rdd.InputFileBlockHolder
-
Sets the thread-local input block.
- set(String, String) - Method in class org.apache.spark.SparkConf
-
Set a configuration variable.
- set(SparkEnv) - Static method in class org.apache.spark.SparkEnv
-
- set(String, String) - Method in class org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(String, boolean) - Method in class org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(String, long) - Method in class org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(long) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Long.
- set(int) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Int.
- set(long, int, int) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given unscaled Long, with a given precision and scale.
- set(BigDecimal, int, int) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given BigDecimal value, with a given precision and scale.
- set(BigDecimal) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given BigDecimal value, inheriting its precision and scale.
- set(BigInteger) - Method in class org.apache.spark.sql.types.Decimal
-
If the value is not in the range of long, convert it to BigDecimal and
the precision and scale are based on the converted value.
- set(Decimal) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Decimal value.
- setAccumulatorId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
int64 accumulator_id = 2;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 44;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- setAccumulatorUpdates(int, StoreTypes.AccumulableInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.AccumulableInfo accumulator_updates = 13;
- setActive(SQLContext) - Static method in class org.apache.spark.sql.SQLContext
-
- setActiveSession(SparkSession) - Static method in class org.apache.spark.sql.SparkSession
-
Changes the SparkSession that will be returned in this thread and its children when
SparkSession.getOrCreate() is called.
- setActiveTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 active_tasks = 9;
- setAddress(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional string address = 1;
- setAddressBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional string address = 1;
- setAddresses(int, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
repeated string addresses = 2;
- setAddTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 add_time = 20;
- setAddTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int64 add_time = 5;
- setAggregationDepth(int) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setAggregationDepth(int) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- setAggregationDepth(int) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregator(Aggregator<K, V, C>) - Method in class org.apache.spark.rdd.ShuffledRDD
-
Set aggregator for RDD's shuffle.
- setAlgo(String) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
Sets Algorithm using a String.
- setAlgo(Enumeration.Value) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setAll(Iterable<Tuple2<String, String>>) - Method in class org.apache.spark.SparkConf
-
Set multiple parameters together
- setAllRemovalsTimeMs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_removals_time_ms = 6;
- setAllUpdatesTimeMs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 all_updates_time_ms = 4;
- setAlpha(double) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setAlpha(Vector) - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for setDocConcentration()
- setAlpha(double) - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for setDocConcentration()
- setAlpha(double) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Sets the constant used in computing confidence in implicit ALS.
- setAmount(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
int64 amount = 2;
- setAmount(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
double amount = 2;
- setAppName(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set the application name.
- setAppName(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setAppName(String) - Method in class org.apache.spark.SparkConf
-
Set a name for your application.
- setAppResource(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set the main application resource.
- setAppResource(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setAppSparkVersion(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string app_spark_version = 8;
- setAppSparkVersionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string app_spark_version = 8;
- setAttempt(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 attempt = 3;
- setAttempt(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 attempt = 3;
- setAttemptId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string attempt_id = 1;
- setAttemptId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 attempt_id = 3;
- setAttemptIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string attempt_id = 1;
- setAttempts(int, StoreTypes.ApplicationAttemptInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- setAttempts(int, StoreTypes.ApplicationAttemptInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ApplicationAttemptInfo attempts = 7;
- setBandwidth(double) - Method in class org.apache.spark.mllib.stat.KernelDensity
-
Sets the bandwidth (standard deviation) of the Gaussian kernel (default: 1.0
).
- setBarrier(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool barrier = 4;
- setBatchDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_duration = 6;
- setBatchId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
int64 batch_id = 5;
- setBeta(double) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setBeta(double) - Method in class org.apache.spark.mllib.clustering.LDA
-
Alias for setTopicConcentration()
- setBinary(boolean) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setBinary(boolean) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
-
- setBinary(boolean) - Method in class org.apache.spark.ml.feature.HashingTF
-
- setBinary(boolean) - Method in class org.apache.spark.mllib.feature.HashingTF
-
If true, term frequency vector will be binary such that non-zero term counts will be set to 1
(default: false)
- setBlacklistedInStages(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 blacklisted_in_stages = 25;
- setBlockName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string block_name = 1;
- setBlockNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string block_name = 1;
- setBlocks(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the number of blocks for both user blocks and product blocks to parallelize the computation
into; pass -1 for an auto-configured number of blocks.
- setBlockSize(int) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param blockSize
.
- setBlockSize(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
Set block size for stacking input data in matrices.
- setBlockSize(int) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
Set block size for stacking input data in matrices.
- setBootstrap(boolean) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setBootstrap(boolean) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setBucketLength(double) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_read = 18;
- setBytesRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double bytes_read = 1;
- setBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 bytes_read = 1;
- setBytesWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double bytes_written = 20;
- setBytesWritten(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double bytes_written = 1;
- setBytesWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 bytes_written = 1;
- setBytesWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 bytes_written = 1;
- setCached(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
bool cached = 3;
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setCacheNodeIds(boolean) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setCallSite(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Pass-through to SparkContext.setCallSite.
- setCallSite(String) - Method in class org.apache.spark.SparkContext
-
Set the thread-local property for overriding the call sites
of actions and RDDs.
- setCallsite(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string callsite = 5;
- setCallsiteBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string callsite = 5;
- setCaseSensitive(boolean) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setCategoricalCols(String[]) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- setCategoricalFeaturesInfo(Map<Integer, Integer>) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
Sets categoricalFeaturesInfo using a Java Map.
- setCategoricalFeaturesInfo(Map<Object, Object>) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setCensorCol(String) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setCheckpointDir(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Set the directory under which RDDs are going to be checkpointed.
- setCheckpointDir(String) - Method in class org.apache.spark.SparkContext
-
Set the directory under which RDDs are going to be checkpointed.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.clustering.LDA
-
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - Method in class org.apache.spark.mllib.clustering.LDA
-
Parameter for set checkpoint interval (greater than or equal to 1) or disable checkpoint (-1).
- setCheckpointInterval(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set period (in iterations) between checkpoints (default = 10).
- setCheckpointInterval(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setChildClusters(int, StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- setChildClusters(int, StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationClusterWrapper child_clusters = 4;
- setChildNodes(int, StoreTypes.RDDOperationNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- setChildNodes(int, StoreTypes.RDDOperationNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationNode child_nodes = 3;
- setClassifier(Classifier<?, ?, ?>) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- setClasspathEntries(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- setClasspathEntries(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings classpath_entries = 6;
- setCluster(StoreTypes.SparkPlanGraphClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- setCluster(StoreTypes.SparkPlanGraphClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper cluster = 2;
- setColdStartStrategy(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setColdStartStrategy(String) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- setCollectSubModels(boolean) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
Whether to collect submodels when fitting.
- setCollectSubModels(boolean) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
Whether to collect submodels when fitting.
- setCommitTimeMs(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 commit_time_ms = 7;
- setCompleted(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
bool completed = 7;
- setCompletedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 completed_tasks = 11;
- setCompletionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 completion_time = 5;
- setCompletionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional int64 completion_time = 9;
- setCompletionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 completion_time = 12;
- setConf(Configuration) - Method in interface org.apache.spark.input.Configurable
-
- setConf(String, String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set a single configuration value for the application.
- setConf(String, String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setConf(Configuration) - Method in class org.apache.spark.ml.image.SamplePathFilter
-
- setConf(Properties) - Method in class org.apache.spark.sql.SQLContext
-
Set Spark SQL configuration properties.
- setConf(String, String) - Method in class org.apache.spark.sql.SQLContext
-
Set the given Spark SQL configuration property.
- setConfig(String, String) - Static method in class org.apache.spark.launcher.SparkLauncher
-
Set a configuration value for the launcher library.
- setConvergenceTol(double) - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Set the largest change in log-likelihood at which convergence is
considered to have occurred.
- setConvergenceTol(double) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the convergence tolerance.
- setConvergenceTol(double) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the convergence tolerance of iterations for L-BFGS.
- setConvergenceTol(double) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the convergence tolerance.
- setCoresGranted(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_granted = 3;
- setCoresPerExecutor(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 cores_per_executor = 5;
- setCorruptMergedBlockChunks(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double corrupt_merged_block_chunks = 1;
- setCorruptMergedBlockChunks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 corrupt_merged_block_chunks = 1;
- setCurrentCatalog(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Sets the current catalog in this session.
- setCurrentDatabase(String) - Method in class org.apache.spark.sql.catalog.Catalog
-
Sets the current database (namespace) in this session.
- setCustomHostname(String) - Static method in class org.apache.spark.util.Utils
-
Allow setting a custom host name because when we run on Mesos we need to use the same
hostname it reports to the master.
- setDAGScheduler(DAGScheduler) - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- setDataDistribution(int, StoreTypes.RDDDataDistribution) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- setDataDistribution(int, StoreTypes.RDDDataDistribution.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDDataDistribution data_distribution = 8;
- setDecayFactor(double) - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the forgetfulness of the previous centroids.
- setDefault(Param<T>, T) - Method in interface org.apache.spark.ml.param.Params
-
Sets a default value for a param.
- setDefault(Seq<ParamPair<?>>) - Method in interface org.apache.spark.ml.param.Params
-
Sets default values for a list of params.
- setDefaultClassLoader(ClassLoader) - Method in class org.apache.spark.serializer.KryoSerializer
-
- setDefaultClassLoader(ClassLoader) - Method in class org.apache.spark.serializer.Serializer
-
Sets a class loader for the serializer to use in deserialization.
- setDefaultSession(SparkSession) - Static method in class org.apache.spark.sql.SparkSession
-
Sets the default SparkSession that is returned by the builder.
- setDegree(int) - Method in class org.apache.spark.ml.feature.PolynomialExpansion
-
- setDelegateCatalog(CatalogPlugin) - Method in interface org.apache.spark.sql.connector.catalog.CatalogExtension
-
- setDelegateCatalog(CatalogPlugin) - Method in class org.apache.spark.sql.connector.catalog.DelegatingCatalogExtension
-
- setDeployMode(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set the deploy mode for the application.
- setDeployMode(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setDesc(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string desc = 3;
- setDesc(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string desc = 3;
- setDescBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string desc = 3;
- setDescBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string desc = 3;
- setDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string description = 3;
- setDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
optional string description = 1;
- setDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string description = 1;
- setDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string description = 3;
- setDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string description = 40;
- setDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string description = 3;
- setDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
optional string description = 1;
- setDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string description = 1;
- setDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string description = 3;
- setDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string description = 40;
- setDeserialized(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool deserialized = 7;
- setDest(long, int, VD, ED) - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- setDetails(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string details = 4;
- setDetails(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string details = 41;
- setDetailsBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string details = 4;
- setDetailsBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string details = 41;
- setDiscoveryScript(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string discoveryScript = 3;
- setDiscoveryScriptBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string discoveryScript = 3;
- setDiskBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double disk_bytes_spilled = 17;
- setDiskBytesSpilled(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double disk_bytes_spilled = 15;
- setDiskBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 disk_bytes_spilled = 14;
- setDiskBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 disk_bytes_spilled = 22;
- setDiskBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 disk_bytes_spilled = 24;
- setDiskBytesSpilled(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double disk_bytes_spilled = 14;
- setDiskBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 disk_bytes_spilled = 9;
- setDiskSize(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 disk_size = 9;
- setDiskUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 disk_used = 6;
- setDiskUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 disk_used = 4;
- setDiskUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 disk_used = 4;
- setDiskUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 disk_used = 7;
- setDistanceMeasure(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setDistanceMeasure(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setDistanceMeasure(String) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setDistanceMeasure(String) - Method in class org.apache.spark.ml.evaluation.ClusteringMetrics
-
- setDistanceMeasure(String) - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Set the distance suite used by the algorithm.
- setDistanceMeasure(String) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the distance suite used by the algorithm.
- setDocConcentration(double[]) - Method in class org.apache.spark.ml.clustering.LDA
-
- setDocConcentration(double) - Method in class org.apache.spark.ml.clustering.LDA
-
- setDocConcentration(Vector) - Method in class org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- setDocConcentration(double) - Method in class org.apache.spark.mllib.clustering.LDA
-
Replicates a Double
docConcentration to create a symmetric prior.
- setDropLast(boolean) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setDropLast(boolean) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setDstCol(String) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 duration = 5;
- setDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double duration = 5;
- setDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 duration = 7;
- setDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 duration = 7;
- setDuration(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double duration = 2;
- setEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- setEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge edges = 2;
- setEdges(int, StoreTypes.SparkPlanGraphEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- setEdges(int, StoreTypes.SparkPlanGraphEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphEdge edges = 3;
- setElasticNetParam(double) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the ElasticNet mixing parameter.
- setElasticNetParam(double) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set the ElasticNet mixing parameter.
- setEndOffset(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string end_offset = 3;
- setEndOffsetBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string end_offset = 3;
- setEndTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 end_time = 3;
- setEndTimestamp(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional int64 end_timestamp = 7;
- setEps(double) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setEpsilon(double) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Sets the value of param epsilon
.
- setEpsilon(double) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the distance threshold within which we've consider centers to have converged.
- setErrorMessage(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string error_message = 10;
- setErrorMessage(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string error_message = 14;
- setErrorMessage(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string error_message = 14;
- setErrorMessageBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string error_message = 10;
- setErrorMessageBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string error_message = 14;
- setErrorMessageBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string error_message = 14;
- setEstimator(Estimator<?>) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setEstimator(Estimator<?>) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- setEstimatorParamMaps(ParamMap[]) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setEstimatorParamMaps(ParamMap[]) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- setEvaluator(Evaluator) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setEvaluator(Evaluator) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- setException(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string exception = 5;
- setExceptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string exception = 5;
- setExcludedInStages(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
repeated int64 excluded_in_stages = 31;
- setExecutionId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
int64 execution_id = 1;
- setExecutionId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 execution_id = 1;
- setExecutorCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_cpu_time = 9;
- setExecutorCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_cpu_time = 17;
- setExecutorCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_cpu_time = 19;
- setExecutorCpuTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_cpu_time = 6;
- setExecutorCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_cpu_time = 4;
- setExecutorDeserializeCpuTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_cpu_time = 7;
- setExecutorDeserializeCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_cpu_time = 15;
- setExecutorDeserializeCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_cpu_time = 17;
- setExecutorDeserializeCpuTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_cpu_time = 4;
- setExecutorDeserializeCpuTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_cpu_time = 2;
- setExecutorDeserializeTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_deserialize_time = 6;
- setExecutorDeserializeTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_deserialize_time = 14;
- setExecutorDeserializeTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_deserialize_time = 16;
- setExecutorDeserializeTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_deserialize_time = 3;
- setExecutorDeserializeTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_deserialize_time = 1;
- setExecutorEnv(String, String) - Method in class org.apache.spark.SparkConf
-
Set an environment variable to be used when launching executors for this application.
- setExecutorEnv(Seq<Tuple2<String, String>>) - Method in class org.apache.spark.SparkConf
-
Set multiple environment variables to be used when launching executors.
- setExecutorEnv(Tuple2<String, String>[]) - Method in class org.apache.spark.SparkConf
-
Set multiple environment variables to be used when launching executors.
- setExecutorId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
optional string executor_id = 3;
- setExecutorId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string executor_id = 2;
- setExecutorId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string executor_id = 8;
- setExecutorId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string executor_id = 8;
- setExecutorIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
optional string executor_id = 3;
- setExecutorIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string executor_id = 2;
- setExecutorIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string executor_id = 8;
- setExecutorIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string executor_id = 8;
- setExecutorMetrics(int, StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- setExecutorMetrics(int, StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated .org.apache.spark.status.protobuf.ExecutorMetrics executor_metrics = 2;
- setExecutorMetricsDistributions(StoreTypes.ExecutorMetricsDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- setExecutorMetricsDistributions(StoreTypes.ExecutorMetricsDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetricsDistributions executor_metrics_distributions = 52;
- setExecutorRunTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double executor_run_time = 8;
- setExecutorRunTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 executor_run_time = 16;
- setExecutorRunTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 executor_run_time = 18;
- setExecutorRunTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double executor_run_time = 5;
- setExecutorRunTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 executor_run_time = 3;
- setExecutors(int, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
repeated string executors = 5;
- setFactorSize(int) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the dimensionality of the factors.
- setFactorSize(int) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the dimensionality of the factors.
- setFailedTasks(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double failed_tasks = 3;
- setFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 failed_tasks = 2;
- setFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 failed_tasks = 10;
- setFailureReason(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string failure_reason = 13;
- setFailureReasonBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string failure_reason = 13;
- setFamily(String) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Sets the value of param family
.
- setFamily(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param family
.
- setFdr(double) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setFdr(double) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setFeatureIndex(int) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setFeatureIndex(int) - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.classification.OneVsRestModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.LDA
-
The features for LDA should be a Vector
representing the word counts in a document.
- setFeaturesCol(String) - Method in class org.apache.spark.ml.clustering.LDAModel
-
The features for LDA should be a Vector
representing the word counts in a document.
- setFeaturesCol(String) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.ChiSqSelectorModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.RFormula
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.PredictionModel
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.Predictor
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setFeaturesCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setFeatureSubsetStrategy(String) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setFeatureSubsetStrategy(String) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setFeatureSubsetStrategy(String) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setFeatureSubsetStrategy(String) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setFeatureType(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setFetchWaitTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double fetch_wait_time = 5;
- setFetchWaitTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 fetch_wait_time = 3;
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- setField(Descriptors.FieldDescriptor, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- setFinalRDDStorageLevel(StorageLevel) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Sets storage level for final RDDs (user/product used in MatrixFactorizationModel).
- setFinalStorageLevel(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setFirstTaskLaunchedTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 first_task_launched_time = 11;
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set whether to fit intercept term.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Whether to fit an intercept term.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Whether to fit an intercept term.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set if we should fit the intercept
Default is true.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set whether to fit intercept term.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets if we should fit the intercept.
- setFitIntercept(boolean) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set if we should fit the intercept.
- setFitLinear(boolean) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set whether to fit linear term.
- setFitLinear(boolean) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set whether to fit linear term.
- setFoldCol(String) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setForceIndexLabel(boolean) - Method in class org.apache.spark.ml.feature.RFormula
-
- setFormula(String) - Method in class org.apache.spark.ml.feature.RFormula
-
Sets the formula to use for this transformer.
- setFpr(double) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setFpr(double) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setFromId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 from_id = 1;
- setFromId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 from_id = 1;
- setFwe(double) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setFwe(double) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setGaps(boolean) - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- setGettingResultTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double getting_result_time = 13;
- setGettingResultTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 getting_result_time = 18;
- setGettingResultTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double getting_result_time = 10;
- setGlobalKrbDebug(boolean) - Static method in class org.apache.spark.util.SecurityUtils
-
- setGradient(Gradient) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the gradient function (of the loss function of one single data example)
to be used for SGD.
- setGradient(Gradient) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the gradient function (of the loss function of one single data example)
to be used for L-BFGS.
- setHadoopProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- setHadoopProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings hadoop_properties = 3;
- setHalfLife(double, String) - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the half life and time unit ("batches" or "points").
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.RFormula
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.VectorAssembler
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.VectorIndexer
-
- setHandleInvalid(String) - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
- setHashAlgorithm(String) - Method in class org.apache.spark.mllib.feature.HashingTF
-
Set the hash algorithm used when mapping term to integer.
- setHasMetrics(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool has_metrics = 15;
- setHost(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string host = 9;
- setHost(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string host = 9;
- setHostBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string host = 9;
- setHostBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string host = 9;
- setHostPort(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string host_port = 2;
- setHostPort(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string host_port = 2;
- setHostPort(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string host_port = 3;
- setHostPortBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string host_port = 2;
- setHostPortBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string host_port = 2;
- setHostPortBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string host_port = 3;
- setId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
int64 id = 1;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string id = 1;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string id = 1;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string id = 1;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string id = 1;
- setId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
int32 id = 1;
- setId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 id = 1;
- setId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
int32 id = 1;
- setId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
int64 id = 1;
- setId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
int64 id = 1;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string id = 2;
- setId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string id = 1;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string id = 1;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string id = 1;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional string id = 1;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string id = 1;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string id = 2;
- setIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string id = 1;
- setIfMissing(String, String) - Method in class org.apache.spark.SparkConf
-
Set a parameter if it isn't already configured
- setImplicitPrefs(boolean) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setImplicitPrefs(boolean) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Sets whether to use implicit preference.
- setImpurity(String) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setImpurity(String) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
The impurity setting is ignored for GBT models.
- setImpurity(String) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setImpurity(String) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setImpurity(String) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
The impurity setting is ignored for GBT models.
- setImpurity(String) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setImpurity(Impurity) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setIncomingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- setIncomingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge incoming_edges = 4;
- setIndex(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 index = 2;
- setIndex(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 index = 2;
- setIndices(int[]) - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- setInfo(StoreTypes.ApplicationEnvironmentInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- setInfo(StoreTypes.ApplicationEnvironmentInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationEnvironmentInfo info = 1;
- setInfo(StoreTypes.ApplicationInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- setInfo(StoreTypes.ApplicationInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.ApplicationInfo info = 1;
- setInfo(StoreTypes.ExecutorStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- setInfo(StoreTypes.ExecutorStageSummary.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorStageSummary info = 4;
- setInfo(StoreTypes.ExecutorSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- setInfo(StoreTypes.ExecutorSummary.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ExecutorSummary info = 1;
- setInfo(StoreTypes.JobData) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- setInfo(StoreTypes.JobData.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
.org.apache.spark.status.protobuf.JobData info = 1;
- setInfo(StoreTypes.ProcessSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- setInfo(StoreTypes.ProcessSummary.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.ProcessSummary info = 1;
- setInfo(StoreTypes.RDDStorageInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- setInfo(StoreTypes.RDDStorageInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDStorageInfo info = 1;
- setInfo(StoreTypes.SpeculationStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- setInfo(StoreTypes.SpeculationStageSummary.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
.org.apache.spark.status.protobuf.SpeculationStageSummary info = 3;
- setInfo(StoreTypes.StageData) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- setInfo(StoreTypes.StageData.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
.org.apache.spark.status.protobuf.StageData info = 1;
- setInitialCenters(Vector[], double[]) - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
-
Specify initial centers directly.
- setInitializationMode(String) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the initialization algorithm.
- setInitializationMode(String) - Method in class org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set the initialization mode.
- setInitializationSteps(int) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the number of steps for the k-means|| initialization mode.
- setInitialModel(LogisticRegressionModel) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- setInitialModel(GaussianMixtureModel) - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Set the initial GMM starting point, bypassing the random initialization.
- setInitialModel(KMeansModel) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the initial starting point, bypassing the random initialization or k-means||
The condition model.k == this.k must be met, failure results
in an IllegalArgumentException.
- setInitialWeights(Vector) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param initialWeights
.
- setInitialWeights(Vector) - Method in class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the initial weights.
- setInitialWeights(Vector) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the initial weights.
- setInitMode(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setInitMode(String) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setInitStd(double) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the standard deviation of initial coefficients.
- setInitStd(double) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the standard deviation of initial coefficients.
- setInitSteps(int) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setInputBytes(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_bytes = 6;
- setInputBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_bytes = 5;
- setInputBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_bytes = 24;
- setInputBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_bytes_read = 26;
- setInputCol(String) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.HashingTF
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.IDF
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.IDFModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.Imputer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.ImputerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.IndexToString
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MaxAbsScaler
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MaxAbsScalerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MinHashLSHModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MinMaxScaler
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.PCA
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.PCAModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.RobustScalerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.StandardScaler
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.StandardScalerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.VectorIndexer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.VectorIndexerModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setInputCol(String) - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- setInputCol(String) - Method in class org.apache.spark.ml.UnaryTransformer
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setInputCols(Seq<String>) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.Imputer
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.ImputerModel
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.Interaction
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- setInputCols(String[]) - Method in class org.apache.spark.ml.feature.VectorAssembler
-
- setInputMetrics(StoreTypes.InputMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- setInputMetrics(StoreTypes.InputMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.InputMetricDistributions input_metrics = 15;
- setInputMetrics(StoreTypes.InputMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- setInputMetrics(StoreTypes.InputMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.InputMetrics input_metrics = 11;
- setInputRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double input_records = 7;
- setInputRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 input_records = 6;
- setInputRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 input_records = 25;
- setInputRecordsRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 input_records_read = 27;
- setInputRowsPerSecond(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double input_rows_per_second = 6;
- setIntercept(boolean) - Method in class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Set if the algorithm should add an intercept.
- setIntermediateRDDStorageLevel(StorageLevel) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Sets storage level for intermediate RDDs (user/product in/out links).
- setIntermediateStorageLevel(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setInterruptOnCancel(boolean) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Set the behavior of job cancellation from jobs started in this thread.
- setInterruptOnCancel(boolean) - Method in class org.apache.spark.SparkContext
-
Set the behavior of job cancellation from jobs started in this thread.
- setInverse(boolean) - Method in class org.apache.spark.ml.feature.DCT
-
- setIsActive(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_active = 3;
- setIsActive(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
bool is_active = 3;
- setIsActive(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
bool is_active = 4;
- setIsBlacklisted(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_blacklisted = 18;
- setIsBlacklistedForStage(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_blacklisted_for_stage = 15;
- setIsExcluded(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
bool is_excluded = 30;
- setIsExcludedForStage(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
bool is_excluded_for_stage = 17;
- setIsotonic(boolean) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setIsotonic(boolean) - Method in class org.apache.spark.mllib.regression.IsotonicRegression
-
Sets the isotonic parameter.
- setIsShufflePushEnabled(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
bool is_shuffle_push_enabled = 63;
- setItemCol(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setItemCol(String) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- setItemsCol(String) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- setItemsCol(String) - Method in class org.apache.spark.ml.fpm.FPGrowthModel
-
- setIterations(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the number of iterations to run.
- setJars(Seq<String>) - Method in class org.apache.spark.SparkConf
-
Set JAR files to distribute to the cluster.
- setJars(String[]) - Method in class org.apache.spark.SparkConf
-
Set JAR files to distribute to the cluster.
- setJavaHome(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
Set a custom JAVA_HOME for launching the Spark application.
- setJavaHome(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_home = 2;
- setJavaHomeBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_home = 2;
- setJavaVersion(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_version = 1;
- setJavaVersionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string java_version = 1;
- setJMapField(Map<K, V>, Function1<Map<K, V>, Object>) - Static method in class org.apache.spark.status.protobuf.Utils
-
- setJobDescription(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Set a human readable description of the current job.
- setJobDescription(String) - Method in class org.apache.spark.SparkContext
-
Set a human readable description of the current job.
- setJobGroup(String, String, boolean) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setJobGroup(String, String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setJobGroup(String, String, boolean) - Method in class org.apache.spark.SparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setJobGroup(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string job_group = 7;
- setJobGroupBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string job_group = 7;
- setJobId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
All IDs are int64 for extendability, even when they are currently int32 in Spark.
- setJobIds(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
repeated int64 job_ids = 2;
- setJobTags(int, String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated string job_tags = 21;
- setJvmGcTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double jvm_gc_time = 11;
- setJvmGcTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 jvm_gc_time = 19;
- setJvmGcTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 jvm_gc_time = 21;
- setJvmGcTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double jvm_gc_time = 8;
- setJvmGcTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 jvm_gc_time = 6;
- setK(int) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setK(int) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setK(int) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setK(int) - Method in class org.apache.spark.ml.clustering.LDA
-
- setK(int) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setK(int) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- setK(int) - Method in class org.apache.spark.ml.feature.PCA
-
- setK(int) - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the desired number of leaf clusters (default: 4).
- setK(int) - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Set the number of Gaussians in the mixture model.
- setK(int) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the number of clusters to create (k).
- setK(int) - Method in class org.apache.spark.mllib.clustering.LDA
-
Set the number of topics to infer, i.e., the number of soft cluster centers.
- setK(int) - Method in class org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set the number of clusters.
- setK(int) - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the number of clusters.
- setKappa(double) - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Learning rate: exponential decay rate---should be between
(0.5, 1.0] to guarantee asymptotic convergence.
- setKeepLastCheckpoint(boolean) - Method in class org.apache.spark.ml.clustering.LDA
-
- setKeepLastCheckpoint(boolean) - Method in class org.apache.spark.mllib.clustering.EMLDAOptimizer
-
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
- setKeyOrdering(Ordering<K>) - Method in class org.apache.spark.rdd.ShuffledRDD
-
Set key ordering for RDD's shuffle.
- setKilledTasks(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double killed_tasks = 5;
- setKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 killed_tasks = 4;
- setLabelCol(String) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- setLabelCol(String) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setLabelCol(String) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setLabelCol(String) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- setLabelCol(String) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- setLabelCol(String) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setLabelCol(String) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setLabelCol(String) - Method in class org.apache.spark.ml.feature.RFormula
-
- setLabelCol(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setLabelCol(String) - Method in class org.apache.spark.ml.Predictor
-
- setLabelCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setLabels(String[]) - Method in class org.apache.spark.ml.feature.IndexToString
-
- setLabelType(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setLambda(double) - Method in class org.apache.spark.mllib.classification.NaiveBayes
-
Set the smoothing parameter.
- setLambda(double) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the regularization parameter, lambda.
- setLastUpdated(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 last_updated = 4;
- setLatestOffset(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string latest_offset = 4;
- setLatestOffsetBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string latest_offset = 4;
- setLatestSeenOffset(Offset) - Method in interface org.apache.spark.sql.connector.read.streaming.AcceptsLatestSeenOffset
-
Callback method to receive the latest seen offset information from streaming execution.
- setLaunchTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 launch_time = 5;
- setLaunchTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 launch_time = 5;
- setLayers(int[]) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param layers
.
- setLeafCol(String) - Method in interface org.apache.spark.ml.tree.DecisionTreeParams
-
- setLearningDecay(double) - Method in class org.apache.spark.ml.clustering.LDA
-
- setLearningOffset(double) - Method in class org.apache.spark.ml.clustering.LDA
-
- setLearningRate(double) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets initial learning rate (default: 0.025).
- setLearningRate(double) - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setLink(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param link
.
- setLinkPower(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param linkPower
.
- setLinkPredictionCol(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the link prediction (linear predictor) column name.
- setLinkPredictionCol(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
Sets the link prediction (linear predictor) column name.
- setLocalBlocksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double local_blocks_fetched = 4;
- setLocalBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_blocks_fetched = 2;
- setLocalBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 local_bytes_read = 6;
- setLocale(String) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setLocalMergedBlocksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_blocks_fetched = 4;
- setLocalMergedBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_blocks_fetched = 4;
- setLocalMergedBytesRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_bytes_read = 8;
- setLocalMergedBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_bytes_read = 8;
- setLocalMergedChunksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double local_merged_chunks_fetched = 6;
- setLocalMergedChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 local_merged_chunks_fetched = 6;
- setLocalProperty(String, String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Set a local property that affects jobs submitted from this thread, and all child
threads, such as the Spark fair scheduler pool.
- setLocalProperty(String, String) - Method in class org.apache.spark.SparkContext
-
Set a local property that affects jobs submitted from this thread, such as the Spark fair
scheduler pool.
- setLogLevel(String) - Method in class org.apache.spark.api.java.JavaSparkContext
-
Control our logLevel.
- setLogLevel(String) - Method in class org.apache.spark.SparkContext
-
Control our logLevel.
- setLogLevel(Level) - Static method in class org.apache.spark.util.Utils
-
configure a new log4j level
- setLoss(String) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Sets the value of param loss
.
- setLoss(Loss) - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setLossType(String) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setLossType(String) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setLower(double) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setLowerBoundsOnCoefficients(Matrix) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the lower bounds on coefficients if fitting under bound constrained optimization.
- setLowerBoundsOnIntercepts(Vector) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the lower bounds on intercepts if fitting under bound constrained optimization.
- setMainClass(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Sets the application class name for Java/Scala applications.
- setMainClass(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setMapSideCombine(boolean) - Method in class org.apache.spark.rdd.ShuffledRDD
-
Set mapSideCombine flag for RDD's shuffle.
- setMaster(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set the Spark master for the application.
- setMaster(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setMaster(String) - Method in class org.apache.spark.SparkConf
-
The master URL to connect to, such as "local" to run locally with one thread, "local[4]" to
run locally with 4 cores, or "spark://master:7077" to run on a Spark standalone cluster.
- setMax(double) - Method in class org.apache.spark.ml.feature.MinMaxScaler
-
- setMax(double) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
-
- setMaxBins(int) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxBins(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMaxBins(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxBins(int) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxBins(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMaxBins(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxBins(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxBlockSizeInMB(double) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Sets the value of param maxBlockSizeInMB
.
- setMaxBlockSizeInMB(double) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Sets the value of param maxBlockSizeInMB
.
- setMaxBlockSizeInMB(double) - Method in class org.apache.spark.ml.clustering.KMeans
-
Sets the value of param maxBlockSizeInMB
.
- setMaxBlockSizeInMB(double) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
Sets the value of param maxBlockSizeInMB
.
- setMaxBlockSizeInMB(double) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Sets the value of param maxBlockSizeInMB
.
- setMaxCategories(int) - Method in class org.apache.spark.ml.feature.VectorIndexer
-
- setMaxCores(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 max_cores = 4;
- setMaxDepth(int) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxDepth(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMaxDepth(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxDepth(int) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxDepth(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMaxDepth(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxDepth(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxDF(double) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setMaxIter(int) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMaxIter(int) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setMaxIter(int) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setMaxIter(int) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setMaxIter(int) - Method in class org.apache.spark.ml.clustering.LDA
-
- setMaxIter(int) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setMaxIter(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setMaxIter(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setMaxIter(int) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the maximum number of iterations.
- setMaxIter(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMaxIter(int) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the maximum number of iterations (applicable for solver "irls").
- setMaxIter(int) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set the maximum number of iterations.
- setMaxIterations(int) - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the max number of k-means iterations to split clusters (default: 20).
- setMaxIterations(int) - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Set the maximum number of iterations allowed.
- setMaxIterations(int) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set maximum number of iterations allowed.
- setMaxIterations(int) - Method in class org.apache.spark.mllib.clustering.LDA
-
Set the maximum number of iterations allowed.
- setMaxIterations(int) - Method in class org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set maximum number of iterations of the power iteration loop
- setMaxLocalProjDBSize(long) - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- setMaxLocalProjDBSize(long) - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Sets the maximum number of items (including delimiters used in the internal storage format)
allowed in a projected database before local processing (default: 32000000L
).
- setMaxMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 max_memory = 19;
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxMemoryInMB(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxPatternLength(int) - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- setMaxPatternLength(int) - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Sets maximal pattern length (default: 10
).
- setMaxSentenceLength(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setMaxSentenceLength(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets the maximum length (in words) of each sentence in the input data.
- setMaxTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 max_tasks = 8;
- setMemoryBytesSpilled(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double memory_bytes_spilled = 16;
- setMemoryBytesSpilled(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double memory_bytes_spilled = 14;
- setMemoryBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 memory_bytes_spilled = 13;
- setMemoryBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 memory_bytes_spilled = 21;
- setMemoryBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 memory_bytes_spilled = 23;
- setMemoryBytesSpilled(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double memory_bytes_spilled = 13;
- setMemoryBytesSpilled(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 memory_bytes_spilled = 8;
- setMemoryMetrics(StoreTypes.MemoryMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- setMemoryMetrics(StoreTypes.MemoryMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.MemoryMetrics memory_metrics = 24;
- setMemoryPerExecutorMb(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional int32 memory_per_executor_mb = 6;
- setMemoryRemaining(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_remaining = 3;
- setMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 memory_used = 5;
- setMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
int64 memory_used = 2;
- setMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
int64 memory_used = 3;
- setMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int64 memory_used = 6;
- setMemoryUsedBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 memory_used_bytes = 8;
- setMemSize(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
int64 mem_size = 8;
- setMergedFetchFallbackCount(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double merged_fetch_fallback_count = 2;
- setMergedFetchFallbackCount(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 merged_fetch_fallback_count = 2;
- setMergerLocs(Seq<BlockManagerId>) - Method in class org.apache.spark.ShuffleDependency
-
- setMetricLabel(double) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setMetricLabel(double) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- setMetricName(String) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- setMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 5;
- setMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- setMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 4;
- setMetrics(int, StoreTypes.SQLPlanMetric) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- setMetrics(int, StoreTypes.SQLPlanMetric.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated .org.apache.spark.status.protobuf.SQLPlanMetric metrics = 7;
- setMetricsProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- setMetricsProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings metrics_properties = 5;
- setMetricType(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string metric_type = 3;
- setMetricTypeBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string metric_type = 3;
- setMetricValuesIsNull(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
bool metric_values_is_null = 13;
- setMin(double) - Method in class org.apache.spark.ml.feature.MinMaxScaler
-
- setMin(double) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
-
- setMinConfidence(double) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- setMinConfidence(double) - Method in class org.apache.spark.ml.fpm.FPGrowthModel
-
- setMinConfidence(double) - Method in class org.apache.spark.mllib.fpm.AssociationRules
-
Sets the minimal confidence (default: 0.8
).
- setMinCount(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setMinCount(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets minCount, the minimum number of times a token must appear to be included in the word2vec
model's vocabulary (default: 5).
- setMinDF(double) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setMinDivisibleClusterSize(double) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setMinDivisibleClusterSize(double) - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the minimum number of points (if greater than or equal to 1.0
) or the minimum proportion
of points (if less than 1.0
) of a divisible cluster (default: 1).
- setMinDocFreq(int) - Method in class org.apache.spark.ml.feature.IDF
-
- setMiniBatchFraction(double) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the mini-batch fraction parameter.
- setMiniBatchFraction(double) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the mini-batch fraction parameter.
- setMiniBatchFraction(double) - Method in class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the fraction of each batch to use for updates.
- setMiniBatchFraction(double) - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Mini-batch fraction in (0, 1], which sets the fraction of document sampled and used in
each iteration.
- setMiniBatchFraction(double) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set fraction of data to be used for each SGD iteration.
- setMiniBatchFraction(double) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the fraction of each batch to use for updates.
- setMinInfoGain(double) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMinInfoGain(double) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMinInfoGain(double) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMinInfoGain(double) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMinInfoGain(double) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMinInfoGain(double) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMinInfoGain(double) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMinInstancesPerNode(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMinSupport(double) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- setMinSupport(double) - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- setMinSupport(double) - Method in class org.apache.spark.mllib.fpm.FPGrowth
-
Sets the minimal support level (default: 0.3
).
- setMinSupport(double) - Method in class org.apache.spark.mllib.fpm.PrefixSpan
-
Sets the minimal support level (default: 0.1
).
- setMinTF(double) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setMinTF(double) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
-
- setMinTokenLength(int) - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setMinWeightFractionPerNode(double) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setMissingValue(double) - Method in class org.apache.spark.ml.feature.Imputer
-
- setModelType(String) - Method in class org.apache.spark.ml.classification.NaiveBayes
-
Set the model type using a string (case-sensitive).
- setModelType(String) - Method in class org.apache.spark.mllib.classification.NaiveBayes
-
Set the model type using a string (case-sensitive).
- setN(int) - Method in class org.apache.spark.ml.feature.NGram
-
- setName(String) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Assign a name to this RDD
- setName(String) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Assign a name to this RDD
- setName(String) - Method in class org.apache.spark.api.java.JavaRDD
-
Assign a name to this RDD
- setName(String) - Method in class org.apache.spark.graphx.impl.EdgeRDDImpl
-
- setName(String) - Method in class org.apache.spark.graphx.impl.VertexRDDImpl
-
- setName(String) - Method in class org.apache.spark.rdd.RDD
-
Assign a name to this RDD
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
optional string name = 1;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
optional string name = 1;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string name = 2;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string name = 1;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string name = 39;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string name = 1;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string name = 1;
- setName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string name = 3;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
optional string name = 1;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
optional string name = 1;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
optional string name = 2;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
optional string name = 1;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string name = 39;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string name = 1;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string name = 1;
- setNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string name = 3;
- setNames(String[]) - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- setNode(StoreTypes.SparkPlanGraphNode) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- setNode(StoreTypes.SparkPlanGraphNode.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
.org.apache.spark.status.protobuf.SparkPlanGraphNode node = 1;
- setNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- setNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 4;
- setNodes(int, StoreTypes.SparkPlanGraphNodeWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- setNodes(int, StoreTypes.SparkPlanGraphNodeWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper nodes = 2;
- setNonnegative(boolean) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setNonnegative(boolean) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set whether the least-squares problems solved at each iteration should have
nonnegativity constraints.
- setNullAt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarArray
-
- setNullAt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatchRow
-
- setNullAt(int) - Method in class org.apache.spark.sql.vectorized.ColumnarRow
-
- setNumActiveStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_stages = 16;
- setNumActiveTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_active_tasks = 10;
- setNumActiveTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_active_tasks = 2;
- setNumActiveTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_active_tasks = 5;
- setNumBins(int) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setNumBlocks(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
Sets both numUserBlocks and numItemBlocks to the specific value.
- setNumBuckets(int) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setNumBucketsArray(int[]) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setNumCachedPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_cached_partitions = 4;
- setNumClasses(int) - Method in class org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
-
Set the number of possible outcomes for k classes classification problem in
Multinomial Logistic Regression.
- setNumClasses(int) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setNumCompletedIndices(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_indices = 15;
- setNumCompletedIndices(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_completed_indices = 9;
- setNumCompletedJobs(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_jobs = 1;
- setNumCompletedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
int32 num_completed_stages = 2;
- setNumCompletedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_stages = 17;
- setNumCompletedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_completed_tasks = 11;
- setNumCompletedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_completed_tasks = 3;
- setNumCompleteTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_complete_tasks = 6;
- setNumCorrections(int) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the number of corrections used in the LBFGS update.
- setNumFailedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_stages = 19;
- setNumFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_failed_tasks = 13;
- setNumFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_failed_tasks = 4;
- setNumFailedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_failed_tasks = 7;
- setNumFeatures(int) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- setNumFeatures(int) - Method in class org.apache.spark.ml.feature.HashingTF
-
- setNumFolds(int) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setNumHashTables(int) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setNumHashTables(int) - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- setNumInputRows(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
int64 num_input_rows = 5;
- setNumItemBlocks(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setNumIterations(int) - Method in class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the number of iterations of gradient descent to run per update.
- setNumIterations(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets number of iterations (default: 1), which should be smaller than or equal to number of
partitions.
- setNumIterations(int) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the number of iterations for SGD.
- setNumIterations(int) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the maximal number of iterations for L-BFGS.
- setNumIterations(int) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the number of iterations of gradient descent to run per update.
- setNumIterations(int) - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setNumKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_killed_tasks = 14;
- setNumKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_killed_tasks = 5;
- setNumKilledTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_killed_tasks = 8;
- setNumOutputRows(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
int64 num_output_rows = 2;
- setNumPartitions(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setNumPartitions(int) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- setNumPartitions(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets number of partitions (default: 1).
- setNumPartitions(int) - Method in class org.apache.spark.mllib.fpm.FPGrowth
-
Sets the number of partitions used by parallel FP-growth (default: same as input data).
- setNumPartitions(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
int32 num_partitions = 3;
- setNumRows(int) - Method in class org.apache.spark.sql.vectorized.ColumnarBatch
-
Sets the number of rows in this batch.
- setNumRowsDroppedByWatermark(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_dropped_by_watermark = 9;
- setNumRowsRemoved(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_removed = 5;
- setNumRowsTotal(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_total = 2;
- setNumRowsUpdated(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_rows_updated = 3;
- setNumShufflePartitions(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_shuffle_partitions = 10;
- setNumSkippedStages(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_stages = 18;
- setNumSkippedTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_skipped_tasks = 12;
- setNumStateStoreInstances(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
int64 num_state_store_instances = 11;
- setNumTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
int32 num_tasks = 9;
- setNumTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
int32 num_tasks = 1;
- setNumTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 num_tasks = 4;
- setNumTopFeatures(int) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setNumTopFeatures(int) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setNumTrees(int) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setNumTrees(int) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setNumUserBlocks(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setOffHeapMemoryRemaining(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_remaining = 8;
- setOffHeapMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 off_heap_memory_used = 6;
- setOffsetCol(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param offsetCol
.
- setOnHeapMemoryRemaining(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_remaining = 7;
- setOnHeapMemoryUsed(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
optional int64 on_heap_memory_used = 5;
- setOperatorName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
optional string operator_name = 1;
- setOperatorNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
optional string operator_name = 1;
- setOptimizeDocConcentration(boolean) - Method in class org.apache.spark.ml.clustering.LDA
-
- setOptimizeDocConcentration(boolean) - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Sets whether to optimize docConcentration parameter during training.
- setOptimizer(String) - Method in class org.apache.spark.ml.clustering.LDA
-
- setOptimizer(LDAOptimizer) - Method in class org.apache.spark.mllib.clustering.LDA
-
LDAOptimizer used to perform the actual calculation (default = EMLDAOptimizer)
- setOptimizer(String) - Method in class org.apache.spark.mllib.clustering.LDA
-
Set the LDAOptimizer used to perform the actual calculation by algorithm name.
- setOrNull(long, int, int) - Method in class org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given unscaled Long, with a given precision and scale,
and return it, or return null if it cannot be set due to overflow.
- setOutgoingEdges(int, StoreTypes.RDDOperationEdge) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- setOutgoingEdges(int, StoreTypes.RDDOperationEdge.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
repeated .org.apache.spark.status.protobuf.RDDOperationEdge outgoing_edges = 3;
- setOutputBytes(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_bytes = 8;
- setOutputBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_bytes = 7;
- setOutputBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_bytes = 26;
- setOutputBytesWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_bytes_written = 28;
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.ChiSqSelectorModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.CountVectorizerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.FeatureHasher
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.HashingTF
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.IDF
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.IDFModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Imputer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.ImputerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.IndexToString
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Interaction
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MaxAbsScaler
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MaxAbsScalerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MinHashLSHModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MinMaxScaler
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.MinMaxScalerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.PCA
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.PCAModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.RobustScalerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.StandardScaler
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.StandardScalerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelectorModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelectorModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VectorAssembler
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VectorIndexer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VectorIndexerModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.VectorSlicer
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setOutputCol(String) - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- setOutputCol(String) - Method in class org.apache.spark.ml.UnaryTransformer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.Imputer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.ImputerModel
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.OneHotEncoder
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.OneHotEncoderModel
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setOutputCols(String[]) - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- setOutputDeterministicLevel(StoreTypes.DeterministicLevel) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- setOutputDeterministicLevelValue(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
.org.apache.spark.status.protobuf.DeterministicLevel output_deterministic_level = 6;
- setOutputMetrics(StoreTypes.OutputMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- setOutputMetrics(StoreTypes.OutputMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.OutputMetricDistributions output_metrics = 16;
- setOutputMetrics(StoreTypes.OutputMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- setOutputMetrics(StoreTypes.OutputMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.OutputMetrics output_metrics = 12;
- setOutputRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double output_records = 9;
- setOutputRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 output_records = 8;
- setOutputRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 output_records = 27;
- setOutputRecordsWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 output_records_written = 29;
- setP(double) - Method in class org.apache.spark.ml.feature.Normalizer
-
- setParallelism(int) - Method in class org.apache.spark.ml.classification.OneVsRest
-
The implementation of parallel one vs.
- setParallelism(int) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
Set the maximum level of parallelism to evaluate models in parallel.
- setParallelism(int) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
Set the maximum level of parallelism to evaluate models in parallel.
- setParent(Estimator<M>) - Method in class org.apache.spark.ml.Model
-
Sets the parent of this model (Java API).
- setPartitionId(Integer) - Method in class org.apache.spark.sql.util.MapperRowCounter
-
- setPartitionId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int32 partition_id = 4;
- setPartitionId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 partition_id = 4;
- setPartitions(int, StoreTypes.RDDPartitionInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- setPartitions(int, StoreTypes.RDDPartitionInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
repeated .org.apache.spark.status.protobuf.RDDPartitionInfo partitions = 9;
- setPathOptionAndCallWithPathParameterError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- setPattern(String) - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- setPeacePeriod(int) - Method in class org.apache.spark.mllib.stat.test.StreamingTest
-
Set the number of initial batches to ignore.
- setPeakExecutionMemory(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double peak_execution_memory = 15;
- setPeakExecutionMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 peak_execution_memory = 23;
- setPeakExecutionMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 peak_execution_memory = 25;
- setPeakExecutionMemory(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double peak_execution_memory = 12;
- setPeakExecutionMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 peak_execution_memory = 10;
- setPeakExecutorMetrics(StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- setPeakExecutorMetrics(StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_executor_metrics = 50;
- setPeakMemoryMetrics(StoreTypes.ExecutorPeakMetricsDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- setPeakMemoryMetrics(StoreTypes.ExecutorPeakMetricsDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
.org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions peak_memory_metrics = 16;
- setPeakMemoryMetrics(StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- setPeakMemoryMetrics(StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 16;
- setPeakMemoryMetrics(StoreTypes.ExecutorMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- setPeakMemoryMetrics(StoreTypes.ExecutorMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional .org.apache.spark.status.protobuf.ExecutorMetrics peak_memory_metrics = 26;
- setPercentile(double) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setPercentile(double) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setPhysicalPlanDescription(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string physical_plan_description = 5;
- setPhysicalPlanDescriptionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
optional string physical_plan_description = 5;
- setPredictionCol(String) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.classification.OneVsRestModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.MultilabelClassificationEvaluator
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.RankingEvaluator
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.fpm.FPGrowth
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.fpm.FPGrowthModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.PredictionModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.Predictor
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setPredictionCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setProbabilityCol(String) - Method in class org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setProbabilityCol(String) - Method in class org.apache.spark.ml.classification.ProbabilisticClassifier
-
- setProbabilityCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setProbabilityCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setProbabilityCol(String) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setProcessedRowsPerSecond(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
double processed_rows_per_second = 7;
- setProductBlocks(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the number of product blocks to parallelize the computation.
- setProgress(StoreTypes.StreamingQueryProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- setProgress(StoreTypes.StreamingQueryProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
.org.apache.spark.status.protobuf.StreamingQueryProgress progress = 1;
- setPropertiesFile(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set a custom properties file with Spark configuration for the application.
- setPropertiesFile(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setProperty(String, String) - Static method in interface org.apache.spark.sql.connector.catalog.NamespaceChange
-
Create a NamespaceChange for setting a namespace property.
- setProperty(String, String) - Static method in interface org.apache.spark.sql.connector.catalog.TableChange
-
Create a TableChange for setting a table property.
- setProperty(String, String) - Static method in interface org.apache.spark.sql.connector.catalog.ViewChange
-
Create a ViewChange for setting a table property.
- setQuantile(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
optional string quantile = 3;
- setQuantileBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
optional string quantile = 3;
- setQuantileCalculationStrategy(Enumeration.Value) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setQuantileProbabilities(double[]) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setQuantileProbabilities(double[]) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setQuantiles(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double quantiles = 1;
- setQuantiles(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
repeated double quantiles = 1;
- setQuantiles(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double quantiles = 1;
- setQuantilesCol(String) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setQuantilesCol(String) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setRandomCenters(int, double, long) - Method in class org.apache.spark.mllib.clustering.StreamingKMeans
-
Initialize random centers, requiring only the number of dimensions.
- setRank(int) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setRank(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the rank of the feature matrices computed (number of features).
- setRatingCol(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setRawPredictionCol(String) - Method in class org.apache.spark.ml.classification.ClassificationModel
-
- setRawPredictionCol(String) - Method in class org.apache.spark.ml.classification.Classifier
-
- setRawPredictionCol(String) - Method in class org.apache.spark.ml.classification.OneVsRest
-
- setRawPredictionCol(String) - Method in class org.apache.spark.ml.classification.OneVsRestModel
-
- setRawPredictionCol(String) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setRddBlocks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 rdd_blocks = 4;
- setRddIds(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
repeated int64 rdd_ids = 43;
- setReadBytes(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_bytes = 1;
- setReadRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double read_records = 2;
- setRecordsRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_read = 19;
- setRecordsRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
repeated double records_read = 2;
- setRecordsRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
int64 records_read = 2;
- setRecordsRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 records_read = 7;
- setRecordsWritten(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double records_written = 21;
- setRecordsWritten(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
repeated double records_written = 2;
- setRecordsWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
int64 records_written = 2;
- setRecordsWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 records_written = 3;
- setRegParam(double) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the L2 regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setRegParam(double) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the L2 regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the regularization parameter for L2 regularization.
- setRegParam(double) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the regularization parameter.
- setRegParam(double) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the regularization parameter.
- setRelativeError(double) - Method in class org.apache.spark.ml.feature.Imputer
-
- setRelativeError(double) - Method in class org.apache.spark.ml.feature.QuantileDiscretizer
-
- setRelativeError(double) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setRemote(String) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Set the Spark master for the application.
- setRemoteBlocksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_blocks_fetched = 3;
- setRemoteBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_blocks_fetched = 1;
- setRemoteBytesRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read = 6;
- setRemoteBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read = 4;
- setRemoteBytesReadToDisk(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_bytes_read_to_disk = 7;
- setRemoteBytesReadToDisk(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_bytes_read_to_disk = 5;
- setRemoteMergedBlocksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_blocks_fetched = 3;
- setRemoteMergedBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_blocks_fetched = 3;
- setRemoteMergedBytesRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_bytes_read = 7;
- setRemoteMergedBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_bytes_read = 7;
- setRemoteMergedChunksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_chunks_fetched = 5;
- setRemoteMergedChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_chunks_fetched = 5;
- setRemoteMergedReqsDuration(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
repeated double remote_merged_reqs_duration = 9;
- setRemoteMergedReqsDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
int64 remote_merged_reqs_duration = 9;
- setRemoteReqsDuration(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double remote_reqs_duration = 9;
- setRemoteReqsDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
int64 remote_reqs_duration = 8;
- setRemoveReason(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string remove_reason = 22;
- setRemoveReasonBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional string remove_reason = 22;
- setRemoveTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
optional int64 remove_time = 21;
- setRemoveTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
optional int64 remove_time = 6;
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- setRepeatedField(Descriptors.FieldDescriptor, int, Object) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- setResourceName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string resource_name = 1;
- setResourceName(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
optional string resource_name = 1;
- setResourceNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string resource_name = 1;
- setResourceNameBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
optional string resource_name = 1;
- setResourceProfileId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 resource_profile_id = 29;
- setResourceProfileId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 resource_profile_id = 49;
- setResourceProfiles(int, StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- setResourceProfiles(int, StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.ResourceProfileInfo resource_profiles = 7;
- setResultFetchStart(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional int64 result_fetch_start = 6;
- setResultFetchStart(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_fetch_start = 6;
- setResultSerializationTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_serialization_time = 12;
- setResultSerializationTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_serialization_time = 20;
- setResultSerializationTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_serialization_time = 22;
- setResultSerializationTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_serialization_time = 9;
- setResultSerializationTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_serialization_time = 7;
- setResultSize(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double result_size = 10;
- setResultSize(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 result_size = 18;
- setResultSize(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 result_size = 20;
- setResultSize(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double result_size = 7;
- setResultSize(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
int64 result_size = 5;
- setRootCluster(StoreTypes.RDDOperationClusterWrapper) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- setRootCluster(StoreTypes.RDDOperationClusterWrapper.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
.org.apache.spark.status.protobuf.RDDOperationClusterWrapper root_cluster = 5;
- setRootExecutionId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 root_execution_id = 2;
- setRpInfo(StoreTypes.ResourceProfileInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- setRpInfo(StoreTypes.ResourceProfileInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
.org.apache.spark.status.protobuf.ResourceProfileInfo rpInfo = 1;
- setRunId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string run_id = 3;
- setRunId(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string run_id = 2;
- setRunIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
optional string run_id = 3;
- setRunIdBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string run_id = 2;
- setRuntime(StoreTypes.RuntimeInfo) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- setRuntime(StoreTypes.RuntimeInfo.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
.org.apache.spark.status.protobuf.RuntimeInfo runtime = 1;
- setSample(RDD<Object>) - Method in class org.apache.spark.mllib.stat.KernelDensity
-
Sets the sample to use for density estimation.
- setSample(JavaRDD<Double>) - Method in class org.apache.spark.mllib.stat.KernelDensity
-
Sets the sample to use for density estimation (for Java users).
- setScalaVersion(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string scala_version = 3;
- setScalaVersionBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
optional string scala_version = 3;
- setScalingVec(Vector) - Method in class org.apache.spark.ml.feature.ElementwiseProduct
-
- setSchedulerDelay(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double scheduler_delay = 14;
- setSchedulerDelay(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 scheduler_delay = 17;
- setSchedulerDelay(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
repeated double scheduler_delay = 11;
- setSchedulingPool(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string scheduling_pool = 42;
- setSchedulingPoolBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional string scheduling_pool = 42;
- setSeed(long) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setSeed(long) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the random seed for weight initialization.
- setSeed(long) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setSeed(long) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the seed for weights initialization if weights are not set
- setSeed(long) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setSeed(long) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
- setSeed(long) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setSeed(long) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setSeed(long) - Method in class org.apache.spark.ml.clustering.LDA
-
- setSeed(long) - Method in class org.apache.spark.ml.clustering.LDAModel
-
- setSeed(long) - Method in class org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setSeed(long) - Method in class org.apache.spark.ml.feature.MinHashLSH
-
- setSeed(long) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setSeed(long) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setSeed(long) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setSeed(long) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the random seed for weight initialization.
- setSeed(long) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setSeed(long) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setSeed(long) - Method in class org.apache.spark.ml.tuning.CrossValidator
-
- setSeed(long) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- setSeed(long) - Method in class org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the random seed (default: hash value of the class name).
- setSeed(long) - Method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Set the random seed
- setSeed(long) - Method in class org.apache.spark.mllib.clustering.KMeans
-
Set the random seed for cluster initialization.
- setSeed(long) - Method in class org.apache.spark.mllib.clustering.LDA
-
Set the random seed for cluster initialization.
- setSeed(long) - Method in class org.apache.spark.mllib.clustering.LocalLDAModel
-
Set the random seed for cluster initialization.
- setSeed(long) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets random seed (default: a random long integer).
- setSeed(long) - Method in class org.apache.spark.mllib.random.ExponentialGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.GammaGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.LogNormalGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.PoissonGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.StandardNormalGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.UniformGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.random.WeibullGenerator
-
- setSeed(long) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Sets a random seed to have deterministic results.
- setSeed(long) - Method in class org.apache.spark.util.random.BernoulliCellSampler
-
- setSeed(long) - Method in class org.apache.spark.util.random.BernoulliSampler
-
- setSeed(long) - Method in class org.apache.spark.util.random.PoissonSampler
-
- setSeed(long) - Method in interface org.apache.spark.util.random.Pseudorandom
-
Set random seed.
- setSelectionMode(String) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setSelectionThreshold(double) - Method in class org.apache.spark.ml.feature.UnivariateFeatureSelector
-
- setSelectorType(String) - Method in class org.apache.spark.ml.feature.ChiSqSelector
-
Deprecated.
- setSelectorType(String) - Method in class org.apache.spark.mllib.feature.ChiSqSelector
-
- setSequenceCol(String) - Method in class org.apache.spark.ml.fpm.PrefixSpan
-
- setSerializer(Serializer) - Method in class org.apache.spark.rdd.CoGroupedRDD
-
Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
- setSerializer(Serializer) - Method in class org.apache.spark.rdd.ShuffledRDD
-
Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
- setShuffleBytesWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_bytes_written = 37;
- setShuffleCorruptMergedBlockChunks(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_corrupt_merged_block_chunks = 33;
- setShuffleCorruptMergedBlockChunks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 53;
- setShuffleCorruptMergedBlockChunks(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_corrupt_merged_block_chunks = 42;
- setShuffleFetchWaitTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_fetch_wait_time = 26;
- setShuffleFetchWaitTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_fetch_wait_time = 30;
- setShuffleFetchWaitTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_fetch_wait_time = 32;
- setShuffleLocalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_local_blocks_fetched = 25;
- setShuffleLocalBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_blocks_fetched = 29;
- setShuffleLocalBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_blocks_fetched = 31;
- setShuffleLocalBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_local_bytes_read = 33;
- setShuffleLocalBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_local_bytes_read = 35;
- setShuffleMergedFetchFallbackCount(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_fetch_fallback_count = 34;
- setShuffleMergedFetchFallbackCount(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_fetch_fallback_count = 54;
- setShuffleMergedFetchFallbackCount(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_fetch_fallback_count = 43;
- setShuffleMergedLocalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_blocks_fetched = 36;
- setShuffleMergedLocalBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_blocks_fetched = 56;
- setShuffleMergedLocalBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_blocks_fetched = 45;
- setShuffleMergedLocalBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_bytes_read = 40;
- setShuffleMergedLocalBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_bytes_read = 60;
- setShuffleMergedLocalBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_bytes_read = 49;
- setShuffleMergedLocalChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_local_chunks_fetched = 38;
- setShuffleMergedLocalChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_local_chunks_fetched = 58;
- setShuffleMergedLocalChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_local_chunks_fetched = 47;
- setShuffleMergedRemoteBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_blocks_fetched = 35;
- setShuffleMergedRemoteBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 55;
- setShuffleMergedRemoteBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_blocks_fetched = 44;
- setShuffleMergedRemoteBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_bytes_read = 39;
- setShuffleMergedRemoteBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_bytes_read = 59;
- setShuffleMergedRemoteBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_bytes_read = 48;
- setShuffleMergedRemoteChunksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_chunks_fetched = 37;
- setShuffleMergedRemoteChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 57;
- setShuffleMergedRemoteChunksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_chunks_fetched = 46;
- setShuffleMergedRemoteReqDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_merged_remote_req_duration = 51;
- setShuffleMergedRemoteReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_merged_remote_reqs_duration = 42;
- setShuffleMergedRemoteReqsDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_merged_remote_reqs_duration = 62;
- setShuffleMergersCount(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int32 shuffle_mergers_count = 64;
- setShufflePushReadMetrics(StoreTypes.ShufflePushReadMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- setShufflePushReadMetrics(StoreTypes.ShufflePushReadMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetrics shuffle_push_read_metrics = 9;
- setShufflePushReadMetricsDist(StoreTypes.ShufflePushReadMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- setShufflePushReadMetricsDist(StoreTypes.ShufflePushReadMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions shuffle_push_read_metrics_dist = 10;
- setShuffleRead(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read = 10;
- setShuffleRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read = 9;
- setShuffleReadBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_read_bytes = 22;
- setShuffleReadBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_bytes = 34;
- setShuffleReadMetrics(StoreTypes.ShuffleReadMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- setShuffleReadMetrics(StoreTypes.ShuffleReadMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetricDistributions shuffle_read_metrics = 17;
- setShuffleReadMetrics(StoreTypes.ShuffleReadMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- setShuffleReadMetrics(StoreTypes.ShuffleReadMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleReadMetrics shuffle_read_metrics = 13;
- setShuffleReadRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_read_records = 11;
- setShuffleReadRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_read_records = 10;
- setShuffleReadRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_read_records = 35;
- setShuffleRecordsRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_records_read = 23;
- setShuffleRecordsRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_read = 36;
- setShuffleRecordsWritten(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_records_written = 39;
- setShuffleRemoteBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_blocks_fetched = 24;
- setShuffleRemoteBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_blocks_fetched = 28;
- setShuffleRemoteBlocksFetched(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_blocks_fetched = 30;
- setShuffleRemoteBytesRead(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read = 27;
- setShuffleRemoteBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read = 31;
- setShuffleRemoteBytesRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read = 33;
- setShuffleRemoteBytesReadToDisk(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_bytes_read_to_disk = 28;
- setShuffleRemoteBytesReadToDisk(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 32;
- setShuffleRemoteBytesReadToDisk(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_bytes_read_to_disk = 34;
- setShuffleRemoteReqsDuration(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_remote_reqs_duration = 41;
- setShuffleRemoteReqsDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_remote_reqs_duration = 61;
- setShuffleRemoteReqsDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_remote_reqs_duration = 50;
- setShuffleTotalBlocksFetched(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_total_blocks_fetched = 29;
- setShuffleWrite(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write = 12;
- setShuffleWrite(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write = 11;
- setShuffleWriteBytes(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_bytes = 30;
- setShuffleWriteBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_bytes = 36;
- setShuffleWriteMetrics(StoreTypes.ShuffleWriteMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- setShuffleWriteMetrics(StoreTypes.ShuffleWriteMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions shuffle_write_metrics = 18;
- setShuffleWriteMetrics(StoreTypes.ShuffleWriteMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- setShuffleWriteMetrics(StoreTypes.ShuffleWriteMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
.org.apache.spark.status.protobuf.ShuffleWriteMetrics shuffle_write_metrics = 14;
- setShuffleWriteRecords(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_records = 31;
- setShuffleWriteRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double shuffle_write_records = 13;
- setShuffleWriteRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 shuffle_write_records = 12;
- setShuffleWriteRecords(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_records = 38;
- setShuffleWriteTime(double) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
double shuffle_write_time = 32;
- setShuffleWriteTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 shuffle_write_time = 37;
- setShuffleWriteTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 shuffle_write_time = 38;
- setSink(StoreTypes.SinkProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- setSink(StoreTypes.SinkProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
.org.apache.spark.status.protobuf.SinkProgress sink = 11;
- setSize(int) - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
- setSkippedStages(int, int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
repeated int32 skipped_stages = 2;
- setSmoothing(double) - Method in class org.apache.spark.ml.classification.NaiveBayes
-
Set the smoothing parameter.
- setSolver(String) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the solver algorithm used for optimization.
- setSolver(String) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param solver
.
- setSolver(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
Sets the value of param solver
.
- setSolver(String) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the solver algorithm used for optimization.
- setSolver(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the solver algorithm used for optimization.
- setSolver(String) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set the solver algorithm used for optimization.
- setSources(int, StoreTypes.SourceProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- setSources(int, StoreTypes.SourceProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.SourceProgress sources = 10;
- setSparkContextSessionConf(SparkSession, Map<Object, Object>) - Static method in class org.apache.spark.sql.api.r.SQLUtils
-
- setSparkHome(String) - Method in class org.apache.spark.launcher.SparkLauncher
-
Set a custom Spark installation location for the application.
- setSparkHome(String) - Method in class org.apache.spark.SparkConf
-
Set the location where Spark is installed on worker nodes.
- setSparkProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- setSparkProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings spark_properties = 2;
- setSparkUser(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string spark_user = 6;
- setSparkUserBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
optional string spark_user = 6;
- setSpeculationSummary(StoreTypes.SpeculationStageSummary) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- setSpeculationSummary(StoreTypes.SpeculationStageSummary.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.SpeculationStageSummary speculation_summary = 47;
- setSpeculative(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
bool speculative = 12;
- setSpeculative(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
bool speculative = 12;
- setSplits(double[]) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setSplitsArray(double[][]) - Method in class org.apache.spark.ml.feature.Bucketizer
-
- setSqlExecutionId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
optional int64 sql_execution_id = 3;
- setSQLReadObject(Function2<DataInputStream, Object, Object>) - Static method in class org.apache.spark.api.r.SerDe
-
- setSQLWriteObject(Function2<DataOutputStream, Object, Object>) - Static method in class org.apache.spark.api.r.SerDe
-
- setSrcCol(String) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setSrcOnly(long, int, VD) - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- setStageAttemptId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int32 stage_attempt_id = 2;
- setStageAttemptId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- setStageAttemptId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int32 stage_attempt_id = 2;
- setStageAttemptId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int32 stage_attempt_id = 41;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 stage_id = 1;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
int64 stage_id = 1;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
int64 stage_id = 1;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
int64 stage_id = 2;
- setStageId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 stage_id = 40;
- setStageIds(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
repeated int64 stage_ids = 6;
- setStageIds(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
repeated int64 stage_ids = 2;
- setStages(PipelineStage[]) - Method in class org.apache.spark.ml.Pipeline
-
- setStages(int, long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
repeated int64 stages = 12;
- setStandardization(boolean) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Whether to standardize the training features before fitting the model.
- setStandardization(boolean) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Whether to standardize the training features before fitting the model.
- setStandardization(boolean) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Whether to standardize the training features before fitting the model.
- setStartOffset(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string start_offset = 2;
- setStartOffsetBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
optional string start_offset = 2;
- setStartTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
int64 start_time = 2;
- setStartTimestamp(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
int64 start_timestamp = 6;
- setStatement(String) - Method in class org.apache.spark.ml.feature.SQLTransformer
-
- setStateOperators(int, StoreTypes.StateOperatorProgress) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- setStateOperators(int, StoreTypes.StateOperatorProgress.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
repeated .org.apache.spark.status.protobuf.StateOperatorProgress state_operators = 9;
- setStatus(StoreTypes.JobExecutionStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- setStatus(StoreTypes.StageStatus) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- setStatus(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string status = 10;
- setStatus(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string status = 10;
- setStatusBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string status = 10;
- setStatusBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string status = 10;
- setStatusValue(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
.org.apache.spark.status.protobuf.JobExecutionStatus status = 8;
- setStatusValue(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
.org.apache.spark.status.protobuf.StageStatus status = 1;
- setStepSize(double) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the initial step size for the first step (like learning rate).
- setStepSize(double) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setStepSize(double) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param stepSize
(applicable only for solver "gd").
- setStepSize(double) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setStepSize(double) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the initial step size for the first step (like learning rate).
- setStepSize(double) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setStepSize(double) - Method in class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the step size for gradient descent.
- setStepSize(double) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the initial step size of SGD for the first step.
- setStepSize(double) - Method in class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the step size for gradient descent.
- setStopWords(String[]) - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
- setStorageLevel(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string storage_level = 2;
- setStorageLevel(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string storage_level = 5;
- setStorageLevel(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string storage_level = 4;
- setStorageLevelBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
optional string storage_level = 2;
- setStorageLevelBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
optional string storage_level = 5;
- setStorageLevelBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
optional string storage_level = 4;
- setStrategy(String) - Method in class org.apache.spark.ml.feature.Imputer
-
Imputation strategy.
- setStringField(String, Function1<String, Object>) - Static method in class org.apache.spark.status.protobuf.Utils
-
- setStringIndexerOrderType(String) - Method in class org.apache.spark.ml.feature.RFormula
-
- setStringOrderType(String) - Method in class org.apache.spark.ml.feature.StringIndexer
-
- setSubmissionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
optional int64 submission_time = 4;
- setSubmissionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
int64 submission_time = 8;
- setSubmissionTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional int64 submission_time = 10;
- setSubsamplingRate(double) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setSubsamplingRate(double) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- setSubsamplingRate(double) - Method in class org.apache.spark.ml.clustering.LDA
-
- setSubsamplingRate(double) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setSubsamplingRate(double) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- setSubsamplingRate(double) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setSucceededTasks(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double succeeded_tasks = 4;
- setSucceededTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int32 succeeded_tasks = 3;
- setSummary(Option<T>) - Method in interface org.apache.spark.ml.util.HasTrainingSummary
-
- setSystemProperties(int, StoreTypes.PairStrings) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- setSystemProperties(int, StoreTypes.PairStrings.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
repeated .org.apache.spark.status.protobuf.PairStrings system_properties = 4;
- setTaskCount(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
int64 task_count = 4;
- setTaskId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
int64 task_id = 1;
- setTaskId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
int64 task_id = 1;
- setTaskLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string task_locality = 11;
- setTaskLocality(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string task_locality = 11;
- setTaskLocalityBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional string task_locality = 11;
- setTaskLocalityBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
optional string task_locality = 11;
- setTaskMetrics(StoreTypes.TaskMetrics) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- setTaskMetrics(StoreTypes.TaskMetrics.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetrics task_metrics = 15;
- setTaskMetricsDistributions(StoreTypes.TaskMetricDistributions) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- setTaskMetricsDistributions(StoreTypes.TaskMetricDistributions.Builder) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
optional .org.apache.spark.status.protobuf.TaskMetricDistributions task_metrics_distributions = 51;
- setTaskTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
repeated double task_time = 2;
- setTaskTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
int64 task_time = 1;
- setTau0(double) - Method in class org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
A (positive) learning parameter that downweights early iterations.
- setTestMethod(String) - Method in class org.apache.spark.mllib.stat.test.StreamingTest
-
Set the statistical method used for significance testing.
- setThreshold(double) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Set threshold in binary classification.
- setThreshold(double) - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
- setThreshold(double) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- setThreshold(double) - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- setThreshold(double) - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
Set threshold in binary classification, in range [0, 1].
- setThreshold(double) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setThreshold(double) - Method in class org.apache.spark.mllib.classification.LogisticRegressionModel
-
Sets the threshold that separates positive predictions from negative predictions
in Binary Logistic Regression.
- setThreshold(double) - Method in class org.apache.spark.mllib.classification.SVMModel
-
Sets the threshold that separates positive predictions from negative predictions.
- setThresholds(double[]) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- setThresholds(double[]) - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- setThresholds(double[]) - Method in interface org.apache.spark.ml.classification.LogisticRegressionParams
-
Set thresholds in multiclass (or binary) classification to adjust the probability of
predicting each class.
- setThresholds(double[]) - Method in class org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setThresholds(double[]) - Method in class org.apache.spark.ml.classification.ProbabilisticClassifier
-
- setThresholds(double[]) - Method in class org.apache.spark.ml.feature.Binarizer
-
- setThroughOrigin(boolean) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setTimeoutDuration(long) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout duration in ms for this key.
- setTimeoutDuration(String) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout duration for this key as a string.
- setTimeoutTimestamp(long) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as milliseconds in epoch time.
- setTimeoutTimestamp(long, String) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as milliseconds in epoch time and an additional
duration as a string (e.g.
- setTimeoutTimestamp(Date) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as a java.sql.Date.
- setTimeoutTimestamp(Date, String) - Method in interface org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as a java.sql.Date and an additional
duration as a string (e.g.
- setTimestamp(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string timestamp = 4;
- setTimestampBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
optional string timestamp = 4;
- setToId(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
int32 to_id = 2;
- setToId(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
int64 to_id = 2;
- setTol(double) - Method in class org.apache.spark.ml.classification.FMClassifier
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setTol(double) - Method in class org.apache.spark.ml.clustering.KMeans
-
- setTol(double) - Method in class org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.regression.FMRegressor
-
Set the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the convergence tolerance of iterations.
- setTol(double) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Set the convergence tolerance of iterations.
- setToLowercase(boolean) - Method in class org.apache.spark.ml.feature.RegexTokenizer
-
- setTopicConcentration(double) - Method in class org.apache.spark.ml.clustering.LDA
-
- setTopicConcentration(double) - Method in class org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics'
distributions over terms.
- setTopicDistributionCol(String) - Method in class org.apache.spark.ml.clustering.LDA
-
- setTopicDistributionCol(String) - Method in class org.apache.spark.ml.clustering.LDAModel
-
- setTotalBlocksFetched(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
repeated double total_blocks_fetched = 8;
- setTotalCores(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_cores = 7;
- setTotalCores(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
int32 total_cores = 4;
- setTotalDuration(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_duration = 13;
- setTotalGcTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_gc_time = 14;
- setTotalInputBytes(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_input_bytes = 15;
- setTotalOffHeapStorageMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_off_heap_storage_memory = 4;
- setTotalOnHeapStorageMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 total_on_heap_storage_memory = 3;
- setTotalShuffleRead(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_read = 16;
- setTotalShuffleWrite(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int64 total_shuffle_write = 17;
- setTotalTasks(int) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
int32 total_tasks = 12;
- setTrainRatio(double) - Method in class org.apache.spark.ml.tuning.TrainValidationSplit
-
- setTreeStrategy(Strategy) - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setUiRoot(ContextHandler, UIRoot) - Static method in class org.apache.spark.status.api.v1.UIRootFromServletContext
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfoWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ApplicationInfoWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.AppSummary.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorPeakMetricsDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.InputMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.OutputMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.PoolData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDDataDistribution.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationClusterWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationEdge.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationNode.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfoWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceInformation.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ResourceProfileWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.RuntimeInfo.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShufflePushReadMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphClusterWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphEdge.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNode.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummary.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.SQLPlanMetric.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics.Builder
-
- setUnknownFields(UnknownFieldSet) - Method in class org.apache.spark.status.protobuf.StoreTypes.TaskResourceRequest.Builder
-
- setUpdate(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string update = 3;
- setUpdateBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string update = 3;
- setUpdater(Updater) - Method in class org.apache.spark.mllib.optimization.GradientDescent
-
Set the updater function to actually perform a gradient step in a given direction.
- setUpdater(Updater) - Method in class org.apache.spark.mllib.optimization.LBFGS
-
Set the updater function to actually perform a gradient step in a given direction.
- SetupDriver(RpcEndpointRef) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver
-
- SetupDriver$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver$
-
- setupGroups(int, org.apache.spark.rdd.DefaultPartitionCoalescer.PartitionLocations) - Method in class org.apache.spark.rdd.DefaultPartitionCoalescer
-
Initializes targetLen partition groups.
- setUpper(double) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setUpperBoundsOnCoefficients(Matrix) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the upper bounds on coefficients if fitting under bound constrained optimization.
- setUpperBoundsOnIntercepts(Vector) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Set the upper bounds on intercepts if fitting under bound constrained optimization.
- setupUI(org.apache.spark.ui.SparkUI) - Method in interface org.apache.spark.status.AppHistoryServerPlugin
-
Sets up UI of this plugin to rebuild the history UI.
- setUsedBins(int) - Method in class org.apache.spark.sql.util.NumericHistogram
-
Set the number of bins currently being used by the histogram.
- setUseDisk(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_disk = 6;
- setUsedOffHeapStorageMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_off_heap_storage_memory = 2;
- setUsedOnHeapStorageMemory(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.MemoryMetrics.Builder
-
int64 used_on_heap_storage_memory = 1;
- setUseMemory(boolean) - Method in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData.Builder
-
bool use_memory = 5;
- setUseNodeIdCache(boolean) - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- setUserBlocks(int) - Method in class org.apache.spark.mllib.recommendation.ALS
-
Set the number of user blocks to parallelize the computation.
- setUserCol(String) - Method in class org.apache.spark.ml.recommendation.ALS
-
- setUserCol(String) - Method in class org.apache.spark.ml.recommendation.ALSModel
-
- setValidateData(boolean) - Method in class org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Set if the algorithm should validate data before training.
- setValidationIndicatorCol(String) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- setValidationIndicatorCol(String) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- setValidationTol(double) - Method in class org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setValue(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string value = 4;
- setValue1(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value1 = 1;
- setValue1Bytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value1 = 1;
- setValue2(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value2 = 2;
- setValue2Bytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.PairStrings.Builder
-
optional string value2 = 2;
- setValueBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.AccumulableInfo.Builder
-
optional string value = 4;
- setVarianceCol(String) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setVarianceCol(String) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setVariancePower(double) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param variancePower
.
- setVarianceThreshold(double) - Method in class org.apache.spark.ml.feature.VarianceThresholdSelector
-
- setVectorSize(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setVectorSize(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets vector size (default: 100).
- setVendor(String) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string vendor = 4;
- setVendorBytes(ByteString) - Method in class org.apache.spark.status.protobuf.StoreTypes.ExecutorResourceRequest.Builder
-
optional string vendor = 4;
- setVerbose(boolean) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Enables verbose reporting for SparkSubmit.
- setVerbose(boolean) - Method in class org.apache.spark.launcher.SparkLauncher
-
- setVocabSize(int) - Method in class org.apache.spark.ml.feature.CountVectorizer
-
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.GBTClassifier
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.LinearSVC
-
Set the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.LogisticRegression
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.NaiveBayes
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.OneVsRest
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.clustering.BisectingKMeans
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.clustering.GaussianMixture
-
- setWeightCol(String) - Method in class org.apache.spark.ml.clustering.KMeans
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- setWeightCol(String) - Method in class org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setWeightCol(String) - Method in class org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setWeightCol(String) - Method in class org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setWeightCol(String) - Method in class org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.GBTRegressor
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param weightCol
.
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.IsotonicRegression
-
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.LinearRegression
-
Whether to over-/under-sample training instances according to the given weights in weightCol.
- setWeightCol(String) - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
Sets the value of param weightCol
.
- setWindowSize(int) - Method in class org.apache.spark.ml.feature.Word2Vec
-
- setWindowSize(int) - Method in class org.apache.spark.mllib.feature.Word2Vec
-
Sets the window of words (default: 5)
- setWindowSize(int) - Method in class org.apache.spark.mllib.stat.test.StreamingTest
-
Set the number of batches to compute significance tests over.
- setWithCentering(boolean) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setWithMean(boolean) - Method in class org.apache.spark.ml.feature.StandardScaler
-
- setWithMean(boolean) - Method in class org.apache.spark.mllib.feature.StandardScalerModel
-
- setWithScaling(boolean) - Method in class org.apache.spark.ml.feature.RobustScaler
-
- setWithStd(boolean) - Method in class org.apache.spark.ml.feature.StandardScaler
-
- setWithStd(boolean) - Method in class org.apache.spark.mllib.feature.StandardScalerModel
-
- setWriteBytes(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_bytes = 1;
- setWriteRecords(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_records = 2;
- setWriteTime(int, double) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetricDistributions.Builder
-
repeated double write_time = 3;
- setWriteTime(long) - Method in class org.apache.spark.status.protobuf.StoreTypes.ShuffleWriteMetrics.Builder
-
int64 write_time = 2;
- sha(Column) - Static method in class org.apache.spark.sql.functions
-
Returns a sha1 hash value as a hex string of the col
.
- sha1(Column) - Static method in class org.apache.spark.sql.functions
-
Calculates the SHA-1 digest of a binary column and returns the value
as a 40 character hex string.
- sha2(Column, int) - Static method in class org.apache.spark.sql.functions
-
Calculates the SHA-2 family of hash functions of a binary column and
returns the value as a hex string.
- shape() - Method in class org.apache.spark.mllib.random.GammaGenerator
-
- SharedParamsCodeGen - Class in org.apache.spark.ml.param.shared
-
Code generator for shared params (sharedParams.scala).
- SharedParamsCodeGen() - Constructor for class org.apache.spark.ml.param.shared.SharedParamsCodeGen
-
- SharedReadWrite$() - Constructor for class org.apache.spark.ml.Pipeline.SharedReadWrite$
-
- sharedState() - Method in class org.apache.spark.sql.SparkSession
-
- shiftLeft(Column, int) - Static method in class org.apache.spark.sql.functions
-
- shiftleft(Column, int) - Static method in class org.apache.spark.sql.functions
-
Shift the given value numBits left.
- shiftRight(Column, int) - Static method in class org.apache.spark.sql.functions
-
- shiftright(Column, int) - Static method in class org.apache.spark.sql.functions
-
(Signed) shift the given value numBits right.
- shiftRightUnsigned(Column, int) - Static method in class org.apache.spark.sql.functions
-
- shiftrightunsigned(Column, int) - Static method in class org.apache.spark.sql.functions
-
Unsigned shift the given value numBits right.
- SHORT() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for nullable short type.
- ShortestPaths - Class in org.apache.spark.graphx.lib
-
Computes shortest paths to the given set of landmark vertices, returning a graph where each
vertex attribute is a map containing the shortest-path distance to each reachable landmark.
- ShortestPaths() - Constructor for class org.apache.spark.graphx.lib.ShortestPaths
-
- ShortExactNumeric - Class in org.apache.spark.sql.types
-
- ShortExactNumeric() - Constructor for class org.apache.spark.sql.types.ShortExactNumeric
-
- shortName() - Method in interface org.apache.spark.ml.util.MLFormatRegister
-
- shortName() - Method in interface org.apache.spark.sql.sources.DataSourceRegister
-
The string that represents the format that this data source provider uses.
- shortTimeUnitString(TimeUnit) - Static method in class org.apache.spark.streaming.ui.UIUtils
-
Return the short string for a TimeUnit
.
- ShortType - Static variable in class org.apache.spark.sql.types.DataTypes
-
Gets the ShortType object.
- ShortType - Class in org.apache.spark.sql.types
-
The data type representing Short
values.
- ShortType() - Constructor for class org.apache.spark.sql.types.ShortType
-
- ShortTypeExpression - Class in org.apache.spark.sql.types
-
- ShortTypeExpression() - Constructor for class org.apache.spark.sql.types.ShortTypeExpression
-
- shortVersion(String) - Static method in class org.apache.spark.util.VersionUtils
-
Given a Spark version string, return the short version string.
- shouldCloseFileAfterWrite(SparkConf, boolean) - Static method in class org.apache.spark.streaming.util.WriteAheadLogUtils
-
- shouldDistributeGaussians(int, int) - Static method in class org.apache.spark.mllib.clustering.GaussianMixture
-
Heuristic to distribute the computation of the MultivariateGaussian
s, approximately when
d is greater than 25 except for when k is very small.
- shouldFilterOutPathName(String) - Static method in class org.apache.spark.util.HadoopFSUtils
-
Checks if we should filter out this path name.
- shouldGoLeft(Vector) - Method in interface org.apache.spark.ml.tree.Split
-
Return true (split to left) or false (split to right).
- shouldGoLeft(int, Split[]) - Method in interface org.apache.spark.ml.tree.Split
-
Return true (split to left) or false (split to right).
- shouldOwn(Param<?>) - Method in interface org.apache.spark.ml.param.Params
-
Validates that the input param belongs to this instance.
- shouldRollover(long) - Method in interface org.apache.spark.util.logging.RollingPolicy
-
Whether rollover should be initiated at this moment
- show(int) - Method in class org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show() - Method in class org.apache.spark.sql.Dataset
-
Displays the top 20 rows of Dataset in a tabular form.
- show(boolean) - Method in class org.apache.spark.sql.Dataset
-
Displays the top 20 rows of Dataset in a tabular form.
- show(int, boolean) - Method in class org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show(int, int) - Method in class org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show(int, int, boolean) - Method in class org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- showBytesDistribution(String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showBytesDistribution(String, Option<org.apache.spark.util.Distribution>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showBytesDistribution(String, org.apache.spark.util.Distribution) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showColumnsNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showColumnsWithConflictDatabasesError(Seq<String>, TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableAsSerdeNotAllowedOnSparkDataSourceTableError(TableIdentifier) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableAsSerdeNotSupportedForV2TablesError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableFailToExecuteUnsupportedConfError(TableIdentifier, StringBuilder) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableFailToExecuteUnsupportedFeatureError(CatalogTable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableNotSupportedOnTempView(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableNotSupportTransactionalHiveTableError(CatalogTable) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showCreateTableOrViewFailToExecuteUnsupportedFeatureError(CatalogTable, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- showDagVizForJob(int, Seq<org.apache.spark.ui.scope.RDDOperationGraph>) - Static method in class org.apache.spark.ui.UIUtils
-
Return a "DAG visualization" DOM element that expands into a visualization for a job.
- showDagVizForStage(int, Option<org.apache.spark.ui.scope.RDDOperationGraph>) - Static method in class org.apache.spark.ui.UIUtils
-
Return a "DAG visualization" DOM element that expands into a visualization for a stage.
- showDistribution(String, org.apache.spark.util.Distribution, Function1<Object, String>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, Option<org.apache.spark.util.Distribution>, Function1<Object, String>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, Option<org.apache.spark.util.Distribution>, String) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showFunctionsInvalidPatternError(String, ParserRuleContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- showFunctionsUnsupportedError(String, SqlBaseParser.IdentifierContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- showMillisDistribution(String, Option<org.apache.spark.util.Distribution>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showMillisDistribution(String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - Static method in class org.apache.spark.scheduler.StatsReportListener
-
- showMillisDistribution(String, Function1<BatchInfo, Option<Object>>) - Method in class org.apache.spark.streaming.scheduler.StatsReportListener
-
- showPartitionNotAllowedOnTableNotPartitionedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- shuffle(Column) - Static method in class org.apache.spark.sql.functions
-
Returns a random permutation of the given array.
- SHUFFLE() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_BATCH() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_BYTES_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_CHUNK() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_CORRUPT_MERGED_BLOCK_CHUNKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_DATA() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_FETCH_WAIT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_FETCH_WAIT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_FETCH_WAIT_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_INDEX() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_LOCAL_BLOCKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_LOCAL_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_LOCAL_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_MERGED_DATA() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_MERGED_FETCH_FALLBACK_COUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_FETCH_FALLBACK_COUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_FETCH_FALLBACK_COUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_INDEX() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_MERGED_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_LOCAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_LOCAL_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_LOCAL_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_LOCAL_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_LOCAL_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_LOCAL_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_LOCAL_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_META() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_MERGED_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_REMOTE_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_REMOTE_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGED_REMOTE_CHUNKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_REMOTE_REQ_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_MERGED_REMOTE_REQS_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_MERGED_REMOTE_REQS_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_MERGERS_COUNT_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_PUSH() - Static method in class org.apache.spark.storage.BlockId
-
- SHUFFLE_PUSH_CORRUPT_MERGED_BLOCK_CHUNKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_FETCH_FALLBACK_COUNT() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_LOCAL_BLOCKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_LOCAL_CHUNKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_LOCAL_READS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_REMOTE_BLOCKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_REMOTE_CHUNKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_REMOTE_READS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_MERGED_REMOTE_REQS_DURATION() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_PUSH_READ_METRICS_DIST_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetricDistributions
-
- SHUFFLE_PUSH_READ_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ShuffleReadMetrics
-
- SHUFFLE_PUSH_READ_METRICS_PREFIX() - Static method in class org.apache.spark.InternalAccumulator
-
- SHUFFLE_READ() - Static method in class org.apache.spark.ui.ToolTips
-
- SHUFFLE_READ_BYTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_READ_BYTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_READ_FETCH_WAIT_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_READ_FETCH_WAIT_TIME() - Static method in class org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SHUFFLE_READ_FETCH_WAIT_TIME() - Static method in class org.apache.spark.ui.ToolTips
-
- SHUFFLE_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- SHUFFLE_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- SHUFFLE_READ_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- SHUFFLE_READ_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- SHUFFLE_READ_METRICS_PREFIX() - Static method in class org.apache.spark.InternalAccumulator
-
- SHUFFLE_READ_RECORDS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_READ_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- SHUFFLE_READ_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- SHUFFLE_READ_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_READ_REMOTE_SIZE() - Static method in class org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SHUFFLE_READ_REMOTE_SIZE() - Static method in class org.apache.spark.ui.ToolTips
-
- SHUFFLE_RECORDS_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_RECORDS_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_RECORDS_WRITTEN_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_REMOTE_BLOCKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_REMOTE_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_REMOTE_BYTES_READ_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_REMOTE_BYTES_READ_TO_DISK_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_REMOTE_BYTES_READ_TO_DISK_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_REMOTE_BYTES_READ_TO_DISK_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_REMOTE_READS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_READS_TO_DISK() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_REQS_DURATION() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_REQS_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_REMOTE_REQS_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_REMOTE_REQS_DURATION_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- SHUFFLE_SERVICE() - Static method in class org.apache.spark.metrics.MetricsSystemInstances
-
- SHUFFLE_TOTAL_BLOCKS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_TOTAL_BLOCKS_FETCHED_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_TOTAL_READS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE() - Static method in class org.apache.spark.ui.ToolTips
-
- SHUFFLE_WRITE_BYTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_WRITE_BYTES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_WRITE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- SHUFFLE_WRITE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- SHUFFLE_WRITE_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetricDistributions
-
- SHUFFLE_WRITE_METRICS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskMetrics
-
- SHUFFLE_WRITE_METRICS_PREFIX() - Static method in class org.apache.spark.InternalAccumulator
-
- SHUFFLE_WRITE_RECORDS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_WRITE_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- SHUFFLE_WRITE_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- SHUFFLE_WRITE_RECORDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_WRITE_SIZE() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE_TIME() - Static method in class org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- SHUFFLE_WRITE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SHUFFLE_WRITE_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- ShuffleBlockBatchId - Class in org.apache.spark.storage
-
- ShuffleBlockBatchId(int, long, int, int) - Constructor for class org.apache.spark.storage.ShuffleBlockBatchId
-
- ShuffleBlockChunkId - Class in org.apache.spark.storage
-
- ShuffleBlockChunkId(int, int, int, int) - Constructor for class org.apache.spark.storage.ShuffleBlockChunkId
-
- ShuffleBlockId - Class in org.apache.spark.storage
-
- ShuffleBlockId(int, long, int) - Constructor for class org.apache.spark.storage.ShuffleBlockId
-
- ShuffleChecksumBlockId - Class in org.apache.spark.storage
-
- ShuffleChecksumBlockId(int, long, int) - Constructor for class org.apache.spark.storage.ShuffleChecksumBlockId
-
- shuffleCleaned(int) - Method in interface org.apache.spark.CleanerListener
-
- shuffleCorruptMergedBlockChunks() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleDataBlockId - Class in org.apache.spark.storage
-
- ShuffleDataBlockId(int, long, int) - Constructor for class org.apache.spark.storage.ShuffleDataBlockId
-
- ShuffleDataIO - Interface in org.apache.spark.shuffle.api
-
:: Private ::
An interface for plugging in modules for storing and reading temporary shuffle data.
- ShuffleDependency<K,V,C> - Class in org.apache.spark
-
:: DeveloperApi ::
Represents a dependency on the output of a shuffle stage.
- ShuffleDependency(RDD<? extends Product2<K, V>>, Partitioner, Serializer, Option<Ordering<K>>, Option<Aggregator<K, V, C>>, boolean, ShuffleWriteProcessor, ClassTag<K>, ClassTag<V>, ClassTag<C>) - Constructor for class org.apache.spark.ShuffleDependency
-
- ShuffledRDD<K,V,C> - Class in org.apache.spark.rdd
-
:: DeveloperApi ::
The resulting RDD from a shuffle (e.g.
- ShuffledRDD(RDD<? extends Product2<K, V>>, Partitioner, ClassTag<K>, ClassTag<V>, ClassTag<C>) - Constructor for class org.apache.spark.rdd.ShuffledRDD
-
- ShuffleDriverComponents - Interface in org.apache.spark.shuffle.api
-
:: Private ::
An interface for building shuffle support modules for the Driver.
- ShuffleExecutorComponents - Interface in org.apache.spark.shuffle.api
-
:: Private ::
An interface for building shuffle support for Executors.
- ShuffleFetchCompletionListener - Class in org.apache.spark.storage
-
A listener to be called at the completion of the ShuffleBlockFetcherIterator
param: data the ShuffleBlockFetcherIterator to process
- ShuffleFetchCompletionListener(ShuffleBlockFetcherIterator) - Constructor for class org.apache.spark.storage.ShuffleFetchCompletionListener
-
- shuffleFetchWaitTime() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleHandle() - Method in class org.apache.spark.ShuffleDependency
-
- shuffleId() - Method in class org.apache.spark.CleanShuffle
-
- shuffleId() - Method in class org.apache.spark.FetchFailed
-
- shuffleId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ShufflePushCompletion
-
- shuffleId() - Method in class org.apache.spark.ShuffleDependency
-
- shuffleId() - Method in class org.apache.spark.storage.BlockManagerMessages.RemoveShuffle
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleBlockBatchId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleBlockChunkId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleChecksumBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleDataBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleIndexBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleMergedBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleMergedDataBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
-
- shuffleId() - Method in class org.apache.spark.storage.ShufflePushBlockId
-
- ShuffleIndexBlockId - Class in org.apache.spark.storage
-
- ShuffleIndexBlockId(int, long, int) - Constructor for class org.apache.spark.storage.ShuffleIndexBlockId
-
- shuffleLocalBlocksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleLocalBytesRead() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleManager() - Method in class org.apache.spark.SparkEnv
-
- ShuffleMapOutputWriter - Interface in org.apache.spark.shuffle.api
-
:: Private ::
A top-level writer that returns child writers for persisting the output of a map task,
and then commits all of the writes as one atomic operation.
- shuffleMergeAllowed() - Method in class org.apache.spark.ShuffleDependency
-
- ShuffleMergedBlockId - Class in org.apache.spark.storage
-
- ShuffleMergedBlockId(int, int, int) - Constructor for class org.apache.spark.storage.ShuffleMergedBlockId
-
- ShuffleMergedDataBlockId - Class in org.apache.spark.storage
-
- ShuffleMergedDataBlockId(String, int, int, int) - Constructor for class org.apache.spark.storage.ShuffleMergedDataBlockId
-
- shuffleMergedFetchFallbackCount() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleMergedIndexBlockId - Class in org.apache.spark.storage
-
- ShuffleMergedIndexBlockId(String, int, int, int) - Constructor for class org.apache.spark.storage.ShuffleMergedIndexBlockId
-
- shuffleMergedLocalBlocksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergedLocalBytesRead() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergedLocalChunksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleMergedMetaBlockId - Class in org.apache.spark.storage
-
- ShuffleMergedMetaBlockId(String, int, int, int) - Constructor for class org.apache.spark.storage.ShuffleMergedMetaBlockId
-
- shuffleMergedRemoteBlocksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergedRemoteBytesRead() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergedRemoteChunksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergedRemoteReqsDuration() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleMergeEnabled() - Method in class org.apache.spark.ShuffleDependency
-
- shuffleMergeFinalized() - Method in class org.apache.spark.ShuffleDependency
-
Returns true if push-based shuffle is disabled or if the shuffle merge for
this shuffle is finalized.
- shuffleMergeId() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ShufflePushCompletion
-
- shuffleMergeId() - Method in class org.apache.spark.ShuffleDependency
-
shuffleMergeId is used to uniquely identify merging process of shuffle
by an indeterminate stage attempt.
- shuffleMergeId() - Method in class org.apache.spark.storage.ShuffleBlockChunkId
-
- shuffleMergeId() - Method in class org.apache.spark.storage.ShuffleMergedBlockId
-
- shuffleMergeId() - Method in class org.apache.spark.storage.ShuffleMergedDataBlockId
-
- shuffleMergeId() - Method in class org.apache.spark.storage.ShuffleMergedIndexBlockId
-
- shuffleMergeId() - Method in class org.apache.spark.storage.ShuffleMergedMetaBlockId
-
- shuffleMergeId() - Method in class org.apache.spark.storage.ShufflePushBlockId
-
- shuffleMergersCount() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleOutputStatus - Interface in org.apache.spark.scheduler
-
A common trait between
MapStatus
and
MergeStatus
.
- ShufflePartitionWriter - Interface in org.apache.spark.shuffle.api
-
:: Private ::
An interface for opening streams to persist partition bytes to a backing data store.
- ShufflePushBlockId - Class in org.apache.spark.storage
-
- ShufflePushBlockId(int, int, int, int) - Constructor for class org.apache.spark.storage.ShufflePushBlockId
-
- ShufflePushCompletion(int, int, int) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ShufflePushCompletion
-
- ShufflePushCompletion$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ShufflePushCompletion$
-
- ShufflePushReadMetricDistributions - Class in org.apache.spark.status.api.v1
-
- ShufflePushReadMetrics - Class in org.apache.spark.status.api.v1
-
- shufflePushReadMetrics() - Method in class org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- shufflePushReadMetricsDist() - Method in class org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- shuffleRead() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
-
- shuffleRead() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleRead$() - Constructor for class org.apache.spark.InternalAccumulator.shuffleRead$
-
- shuffleReadBytes() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleReadMetricDistributions - Class in org.apache.spark.status.api.v1
-
- ShuffleReadMetrics - Class in org.apache.spark.status.api.v1
-
- shuffleReadMetrics() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
-
- shuffleReadMetrics() - Method in class org.apache.spark.status.api.v1.TaskMetrics
-
- shuffleReadRecords() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
-
- shuffleReadRecords() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleReadRecords() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleRemoteBlocksFetched() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleRemoteBytesRead() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleRemoteBytesReadToDisk() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleRemoteReqsDuration() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleStatus - Class in org.apache.spark
-
Helper class used by the MapOutputTrackerMaster
to perform bookkeeping for a single
ShuffleMapStage.
- ShuffleStatus(int, int) - Constructor for class org.apache.spark.ShuffleStatus
-
- shuffleWrite() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
-
- shuffleWrite() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleWrite$() - Constructor for class org.apache.spark.InternalAccumulator.shuffleWrite$
-
- shuffleWriteBytes() - Method in class org.apache.spark.status.api.v1.StageData
-
- ShuffleWriteMetricDistributions - Class in org.apache.spark.status.api.v1
-
- ShuffleWriteMetrics - Class in org.apache.spark.status.api.v1
-
- shuffleWriteMetrics() - Method in class org.apache.spark.status.api.v1.TaskMetricDistributions
-
- shuffleWriteMetrics() - Method in class org.apache.spark.status.api.v1.TaskMetrics
-
- shuffleWriteRecords() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
-
- shuffleWriteRecords() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleWriteRecords() - Method in class org.apache.spark.status.api.v1.StageData
-
- shuffleWriterProcessor() - Method in class org.apache.spark.ShuffleDependency
-
- shuffleWriteTime() - Method in class org.apache.spark.status.api.v1.StageData
-
- shutdown() - Method in interface org.apache.spark.api.plugin.DriverPlugin
-
Informs the plugin that the Spark application is shutting down.
- shutdown() - Method in interface org.apache.spark.api.plugin.ExecutorPlugin
-
Clean up and terminate this plugin.
- Shutdown(int) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.Shutdown
-
- shutdown(ExecutorService, Duration) - Static method in class org.apache.spark.util.ThreadUtils
-
- Shutdown$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.Shutdown$
-
- ShutdownHookManager - Class in org.apache.spark.util
-
Various utility methods used by Spark.
- ShutdownHookManager() - Constructor for class org.apache.spark.util.ShutdownHookManager
-
- sigma() - Method in class org.apache.spark.ml.classification.NaiveBayesModel
-
- sigma() - Method in class org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
- sigmas() - Method in class org.apache.spark.mllib.clustering.ExpectationSum
-
- sign(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the signum of the given value.
- SignalUtils - Class in org.apache.spark.util
-
Contains utilities for working with posix signals.
- SignalUtils() - Constructor for class org.apache.spark.util.SignalUtils
-
- signum(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the signum of the given value.
- signum(String) - Static method in class org.apache.spark.sql.functions
-
Computes the signum of the given column.
- signum(T) - Static method in class org.apache.spark.sql.types.ByteExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.DecimalExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.DoubleExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.FloatExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.IntegerExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.LongExactNumeric
-
- signum(T) - Static method in class org.apache.spark.sql.types.ShortExactNumeric
-
- silhouette() - Method in class org.apache.spark.ml.evaluation.ClusteringMetrics
-
Returns the silhouette score
- SimpleFutureAction<T> - Class in org.apache.spark
-
A
FutureAction
holding the result of an action that triggers a single job.
- SimpleMetricsCachedBatch - Interface in org.apache.spark.sql.columnar
-
- SimpleMetricsCachedBatchSerializer - Class in org.apache.spark.sql.columnar
-
- SimpleMetricsCachedBatchSerializer() - Constructor for class org.apache.spark.sql.columnar.SimpleMetricsCachedBatchSerializer
-
- simpleString() - Method in class org.apache.spark.sql.types.ArrayType
-
- simpleString() - Static method in class org.apache.spark.sql.types.BinaryType
-
- simpleString() - Static method in class org.apache.spark.sql.types.BooleanType
-
- simpleString() - Method in class org.apache.spark.sql.types.ByteType
-
- simpleString() - Static method in class org.apache.spark.sql.types.CalendarIntervalType
-
- simpleString() - Method in class org.apache.spark.sql.types.DataType
-
Readable string representation for the type.
- simpleString() - Static method in class org.apache.spark.sql.types.DateType
-
- simpleString() - Method in class org.apache.spark.sql.types.DecimalType
-
- simpleString() - Static method in class org.apache.spark.sql.types.DoubleType
-
- simpleString() - Static method in class org.apache.spark.sql.types.FloatType
-
- simpleString() - Method in class org.apache.spark.sql.types.IntegerType
-
- simpleString() - Method in class org.apache.spark.sql.types.LongType
-
- simpleString() - Method in class org.apache.spark.sql.types.MapType
-
- simpleString() - Static method in class org.apache.spark.sql.types.NullType
-
- simpleString() - Method in class org.apache.spark.sql.types.ObjectType
-
- simpleString() - Method in class org.apache.spark.sql.types.ShortType
-
- simpleString() - Static method in class org.apache.spark.sql.types.StringType
-
- simpleString() - Method in class org.apache.spark.sql.types.StructType
-
- simpleString() - Static method in class org.apache.spark.sql.types.TimestampNTZType
-
- simpleString() - Static method in class org.apache.spark.sql.types.TimestampType
-
- SimpleUpdater - Class in org.apache.spark.mllib.optimization
-
A simple updater for gradient descent *without* any regularization.
- SimpleUpdater() - Constructor for class org.apache.spark.mllib.optimization.SimpleUpdater
-
- sin(Column) - Static method in class org.apache.spark.sql.functions
-
- sin(String) - Static method in class org.apache.spark.sql.functions
-
- SingleSpillShuffleMapOutputWriter - Interface in org.apache.spark.shuffle.api
-
Optional extension for partition writing that is optimized for transferring a single
file to the backing store.
- singleTableStarInCountNotAllowedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- SingleValueExecutorMetricType - Interface in org.apache.spark.metrics
-
- SingularValueDecomposition<UType,VType> - Class in org.apache.spark.mllib.linalg
-
Represents singular value decomposition (SVD) factors.
- SingularValueDecomposition(UType, Vector, VType) - Constructor for class org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- sinh(Column) - Static method in class org.apache.spark.sql.functions
-
- sinh(String) - Static method in class org.apache.spark.sql.functions
-
- Sink - Interface in org.apache.spark.metrics.sink
-
- sink() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
-
- SINK_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- SinkProgress - Class in org.apache.spark.sql.streaming
-
Information about progress made for a sink in the execution of a
StreamingQuery
during a trigger.
- SinkProgressSerializer - Class in org.apache.spark.status.protobuf.sql
-
- SinkProgressSerializer() - Constructor for class org.apache.spark.status.protobuf.sql.SinkProgressSerializer
-
- size() - Method in class org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- size() - Method in class org.apache.spark.ml.attribute.AttributeGroup
-
Size of the attribute group.
- size() - Method in class org.apache.spark.ml.feature.VectorSizeHint
-
The size of Vectors in inputCol
.
- size() - Method in class org.apache.spark.ml.linalg.DenseVector
-
- size() - Method in class org.apache.spark.ml.linalg.SparseVector
-
- size() - Method in interface org.apache.spark.ml.linalg.Vector
-
Size of the vector.
- size() - Method in class org.apache.spark.ml.param.ParamMap
-
Number of param pairs in this map.
- size() - Method in class org.apache.spark.mllib.linalg.DenseVector
-
- size() - Method in class org.apache.spark.mllib.linalg.SparseVector
-
- size() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Size of the vector.
- size(Column) - Static method in class org.apache.spark.sql.functions
-
Returns length of array or map.
- size() - Method in interface org.apache.spark.sql.Row
-
Number of elements in the Row.
- size() - Method in class org.apache.spark.sql.util.CaseInsensitiveStringMap
-
- size(KVStoreView<T>) - Static method in class org.apache.spark.status.KVUtils
-
- size() - Method in interface org.apache.spark.storage.BlockData
-
- size() - Method in class org.apache.spark.storage.DiskBlockData
-
- size() - Method in class org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- size() - Method in interface org.apache.spark.storage.memory.MemoryEntry
-
- size() - Method in class org.apache.spark.storage.memory.SerializedMemoryEntry
-
- size(Iterator<?>) - Static method in class org.apache.spark.util.Iterators
-
Counts the number of elements of an iterator using a while loop rather than calling
TraversableOnce.size()
because it uses a for loop, which is slightly slower
in the current version of Scala.
- SIZE_IN_MEMORY() - Static method in class org.apache.spark.ui.storage.ToolTips
-
- SIZE_ON_DISK() - Static method in class org.apache.spark.ui.storage.ToolTips
-
- SizeEstimator - Class in org.apache.spark.util
-
:: DeveloperApi ::
Estimates the sizes of Java objects (number of bytes of memory they occupy), for use in
memory-aware caches.
- SizeEstimator() - Constructor for class org.apache.spark.util.SizeEstimator
-
- sizeInBytes() - Method in interface org.apache.spark.sql.columnar.CachedBatch
-
- sizeInBytes() - Method in interface org.apache.spark.sql.columnar.SimpleMetricsCachedBatch
-
- sizeInBytes() - Method in interface org.apache.spark.sql.connector.read.Statistics
-
- sizeInBytes() - Method in class org.apache.spark.sql.sources.BaseRelation
-
Returns an estimated size of this relation in bytes.
- sketch(RDD<K>, int, ClassTag<K>) - Static method in class org.apache.spark.RangePartitioner
-
Sketches the input RDD via reservoir sampling on each partition.
- skewness(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the skewness of the values in a group.
- skewness(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the skewness of the values in a group.
- skip(long) - Method in class org.apache.spark.io.NioBufferedFileInputStream
-
- skip(long) - Method in class org.apache.spark.io.ReadAheadInputStream
-
- skip(long) - Method in class org.apache.spark.storage.BufferReleasingInputStream
-
- SKIPPED_STAGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- skippedStages() - Method in class org.apache.spark.status.LiveJob
-
- skippedTasks() - Method in class org.apache.spark.status.LiveJob
-
- skipWhitespace() - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- slice(Column, int, int) - Static method in class org.apache.spark.sql.functions
-
Returns an array containing all the elements in x
from index start
(or starting from the
end if start
is negative) with the specified length
.
- slice(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns an array containing all the elements in x
from index start
(or starting from the
end if start
is negative) with the specified length
.
- slice(Time, Time) - Method in interface org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return all the RDDs between 'fromDuration' to 'toDuration' (both included)
- slice(org.apache.spark.streaming.Interval) - Method in class org.apache.spark.streaming.dstream.DStream
-
Return all the RDDs defined by the Interval object (both end times included)
- slice(Time, Time) - Method in class org.apache.spark.streaming.dstream.DStream
-
Return all the RDDs between 'fromTime' to 'toTime' (both included)
- slideDuration() - Method in class org.apache.spark.streaming.dstream.DStream
-
Time interval after which the DStream generates an RDD
- slideDuration() - Method in class org.apache.spark.streaming.dstream.InputDStream
-
- sliding(int, int) - Method in class org.apache.spark.mllib.rdd.RDDFunctions
-
Returns an RDD from grouping items of its parent RDD in fixed size blocks by passing a sliding
window over them.
- sliding(int) - Method in class org.apache.spark.mllib.rdd.RDDFunctions
-
sliding(Int, Int)*
with step = 1.
- slotsPerAddress() - Method in interface org.apache.spark.resource.ResourceAllocator
-
- smoothing() - Method in class org.apache.spark.ml.classification.NaiveBayes
-
- smoothing() - Method in class org.apache.spark.ml.classification.NaiveBayesModel
-
- smoothing() - Method in interface org.apache.spark.ml.classification.NaiveBayesParams
-
The smoothing parameter.
- SnappyCompressionCodec - Class in org.apache.spark.io
-
- SnappyCompressionCodec(SparkConf) - Constructor for class org.apache.spark.io.SnappyCompressionCodec
-
- socketStream(String, int, Function<InputStream, Iterable<T>>, StorageLevel) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Create an input stream from network source hostname:port.
- socketStream(String, int, Function1<InputStream, Iterator<T>>, StorageLevel, ClassTag<T>) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Creates an input stream from TCP source hostname:port.
- socketTextStream(String, int, StorageLevel) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Create an input stream from network source hostname:port.
- socketTextStream(String, int) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Create an input stream from network source hostname:port.
- socketTextStream(String, int, StorageLevel) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Creates an input stream from TCP source hostname:port.
- softmax(double[]) - Static method in class org.apache.spark.ml.impl.Utils
-
Perform in-place softmax conversion.
- softmax(double[], int, int, int, double[]) - Static method in class org.apache.spark.ml.impl.Utils
-
Perform softmax conversion.
- solve(double, double, DenseVector, DenseVector, DenseVector) - Method in interface org.apache.spark.ml.optim.NormalEquationSolver
-
Solve the normal equations from summary statistics.
- solve(ALS.NormalEquation, double) - Method in interface org.apache.spark.ml.recommendation.ALS.LeastSquaresNESolver
-
Solves a least squares problem with regularization (possibly with other constraints).
- solve(double[], double[]) - Static method in class org.apache.spark.mllib.linalg.CholeskyDecomposition
-
Solves a symmetric positive definite linear system via Cholesky factorization.
- solve(double[], double[], NNLS.Workspace) - Static method in class org.apache.spark.mllib.optimization.NNLS
-
Solve a least squares problem, possibly with nonnegativity constraints, by a modified
projected gradient method.
- solver() - Method in class org.apache.spark.ml.classification.FMClassificationModel
-
- solver() - Method in class org.apache.spark.ml.classification.FMClassifier
-
- solver() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- solver() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- solver() - Method in interface org.apache.spark.ml.classification.MultilayerPerceptronParams
-
The solver algorithm for optimization.
- solver() - Method in class org.apache.spark.ml.clustering.KMeans
-
- solver() - Method in class org.apache.spark.ml.clustering.KMeansModel
-
- solver() - Method in interface org.apache.spark.ml.clustering.KMeansParams
-
Param for the name of optimization method used in KMeans.
- solver() - Method in interface org.apache.spark.ml.param.shared.HasSolver
-
Param for the solver algorithm for optimization.
- solver() - Method in interface org.apache.spark.ml.regression.FactorizationMachinesParams
-
The solver algorithm for optimization.
- solver() - Method in class org.apache.spark.ml.regression.FMRegressionModel
-
- solver() - Method in class org.apache.spark.ml.regression.FMRegressor
-
- solver() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- solver() - Method in interface org.apache.spark.ml.regression.GeneralizedLinearRegressionBase
-
The solver algorithm for optimization.
- solver() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- solver() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
- solver() - Method in class org.apache.spark.ml.regression.LinearRegression
-
- solver() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- solver() - Method in interface org.apache.spark.ml.regression.LinearRegressionParams
-
The solver algorithm for optimization.
- some(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns true if at least one value of e
is true.
- Sort() - Static method in class org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- sort(Expression, SortDirection, NullOrdering) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
-
Create a sort expression.
- sort(Expression, SortDirection) - Static method in class org.apache.spark.sql.connector.expressions.Expressions
-
Create a sort expression.
- sort(Expression, SortDirection, NullOrdering) - Static method in class org.apache.spark.sql.connector.expressions.LogicalExpressions
-
- sort(String, String...) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the specified column, all in ascending order.
- sort(Column...) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the given expressions.
- sort(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the specified column, all in ascending order.
- sort(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the given expressions.
- sort_array(Column) - Static method in class org.apache.spark.sql.functions
-
Sorts the input array for the given column in ascending order,
according to the natural ordering of the array elements.
- sort_array(Column, boolean) - Static method in class org.apache.spark.sql.functions
-
Sorts the input array for the given column in ascending or descending order,
according to the natural ordering of the array elements.
- sortBy(Function<T, S>, boolean, int) - Method in class org.apache.spark.api.java.JavaRDD
-
Return this RDD sorted by the given key function.
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - Method in class org.apache.spark.rdd.RDD
-
Return this RDD sorted by the given key function.
- sortBy(String, String...) - Method in class org.apache.spark.sql.DataFrameWriter
-
Sorts the output in each bucket by the given columns.
- sortBy(String, Seq<String>) - Method in class org.apache.spark.sql.DataFrameWriter
-
Sorts the output in each bucket by the given columns.
- sortByKey() - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements in
ascending order.
- sortByKey(boolean) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(boolean, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>, boolean) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>, boolean, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(boolean, int) - Method in class org.apache.spark.rdd.OrderedRDDFunctions
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByWithoutBucketingError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- SortDirection - Enum in org.apache.spark.sql.connector.expressions
-
A sort direction used in sorting expressions.
- SortOrder - Interface in org.apache.spark.sql.connector.expressions
-
Represents a sort order in the public expression API.
- sortWithinPartitions(String, String...) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(Column...) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(String, Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(Seq<Column>) - Method in class org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- soundex(Column) - Static method in class org.apache.spark.sql.functions
-
Returns the soundex code for the specified expression.
- Source - Interface in org.apache.spark.metrics.source
-
- source() - Method in class org.apache.spark.sql.connector.expressions.Extract
-
- SOURCE_NAME_CONSOLE() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCE_NAME_FOREACH() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCE_NAME_FOREACH_BATCH() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCE_NAME_MEMORY() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCE_NAME_NOOP() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCE_NAME_TABLE() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- sourceName() - Static method in class org.apache.spark.metrics.source.CodegenMetrics
-
- sourceName() - Static method in class org.apache.spark.metrics.source.HiveCatalogMetrics
-
- sourceName() - Method in interface org.apache.spark.metrics.source.Source
-
- sourceNotSupportedWithContinuousTriggerError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- SourceProgress - Class in org.apache.spark.sql.streaming
-
Information about progress made for a source in the execution of a
StreamingQuery
during a trigger.
- SourceProgressSerializer - Class in org.apache.spark.status.protobuf.sql
-
- SourceProgressSerializer() - Constructor for class org.apache.spark.status.protobuf.sql.SourceProgressSerializer
-
- sources() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
-
- SOURCES_ALLOW_ONE_TIME_QUERY() - Static method in class org.apache.spark.sql.streaming.DataStreamWriter
-
- SOURCES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- sourceSchema(SQLContext, Option<StructType>, String, Map<String, String>) - Method in interface org.apache.spark.sql.sources.StreamSourceProvider
-
Returns the name and schema of the source that can be used to continually read data.
- spark() - Method in class org.apache.spark.status.api.v1.VersionInfo
-
- spark_branch() - Static method in class org.apache.spark.SparkBuildInfo
-
- spark_build_date() - Static method in class org.apache.spark.SparkBuildInfo
-
- spark_build_user() - Static method in class org.apache.spark.SparkBuildInfo
-
- SPARK_CONNECTOR_NAME() - Static method in class org.apache.spark.ui.JettyUtils
-
- SPARK_CONTEXT_SHUTDOWN_PRIORITY() - Static method in class org.apache.spark.util.ShutdownHookManager
-
The shutdown priority of the SparkContext instance.
- spark_doc_root() - Static method in class org.apache.spark.SparkBuildInfo
-
- SPARK_IO_ENCRYPTION_COMMONS_CONFIG_PREFIX() - Static method in class org.apache.spark.security.CryptoStreamUtils
-
- SPARK_LOCAL_REMOTE - Static variable in class org.apache.spark.launcher.SparkLauncher
-
- SPARK_MASTER - Static variable in class org.apache.spark.launcher.SparkLauncher
-
The Spark master.
- spark_partition_id() - Static method in class org.apache.spark.sql.functions
-
Partition ID.
- SPARK_PROPERTIES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- SPARK_REGEX() - Static method in class org.apache.spark.SparkMasterRegex
-
- SPARK_REMOTE - Static variable in class org.apache.spark.launcher.SparkLauncher
-
The Spark remote.
- spark_repo_url() - Static method in class org.apache.spark.SparkBuildInfo
-
- spark_revision() - Static method in class org.apache.spark.SparkBuildInfo
-
- SPARK_USER_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- spark_version() - Static method in class org.apache.spark.SparkBuildInfo
-
- SparkAppConfig(Seq<Tuple2<String, String>>, Option<byte[]>, Option<byte[]>, ResourceProfile) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig
-
- SparkAppConfig$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig$
-
- SparkAppHandle - Interface in org.apache.spark.launcher
-
A handle to a running Spark application.
- SparkAppHandle.Listener - Interface in org.apache.spark.launcher
-
Listener for updates to a handle's state.
- SparkAppHandle.State - Enum in org.apache.spark.launcher
-
Represents the application's state.
- SparkAWSCredentials - Interface in org.apache.spark.streaming.kinesis
-
Serializable interface providing a method executors can call to obtain an
AWSCredentialsProvider instance for authenticating to AWS services.
- SparkAWSCredentials.Builder - Class in org.apache.spark.streaming.kinesis
-
- SparkBuildInfo - Class in org.apache.spark
-
- SparkBuildInfo() - Constructor for class org.apache.spark.SparkBuildInfo
-
- SparkClassUtils - Interface in org.apache.spark.util
-
- SparkConf - Class in org.apache.spark
-
Configuration for a Spark application.
- SparkConf(boolean) - Constructor for class org.apache.spark.SparkConf
-
- SparkConf() - Constructor for class org.apache.spark.SparkConf
-
Create a SparkConf that loads defaults from system properties and the classpath
- sparkContext() - Method in class org.apache.spark.rdd.RDD
-
The SparkContext that created this RDD.
- SparkContext - Class in org.apache.spark
-
Main entry point for Spark functionality.
- SparkContext(SparkConf) - Constructor for class org.apache.spark.SparkContext
-
- SparkContext() - Constructor for class org.apache.spark.SparkContext
-
Create a SparkContext that loads settings from system properties (for instance, when
launching with ./bin/spark-submit).
- SparkContext(String, String, SparkConf) - Constructor for class org.apache.spark.SparkContext
-
Alternative constructor that allows setting common Spark properties directly
- SparkContext(String, String, String, Seq<String>, Map<String, String>) - Constructor for class org.apache.spark.SparkContext
-
Alternative constructor that allows setting common Spark properties directly
- sparkContext() - Method in class org.apache.spark.sql.SparkSession
-
- sparkContext() - Method in class org.apache.spark.sql.SQLContext
-
- sparkContext() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
The underlying SparkContext
- sparkContext() - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Return the associated Spark context
- SparkCoreErrors - Class in org.apache.spark.errors
-
Object for grouping error messages from (most) exceptions thrown during query execution.
- SparkCoreErrors() - Constructor for class org.apache.spark.errors.SparkCoreErrors
-
- SparkDataStream - Interface in org.apache.spark.sql.connector.read.streaming
-
The base interface representing a readable data stream in a Spark streaming query.
- SparkEnv - Class in org.apache.spark
-
:: DeveloperApi ::
Holds all the runtime environment objects for a running Spark instance (either master or worker),
including the serializer, RpcEnv, block manager, map output tracker, etc.
- SparkEnv(String, RpcEnv, Serializer, Serializer, org.apache.spark.serializer.SerializerManager, MapOutputTracker, ShuffleManager, org.apache.spark.broadcast.BroadcastManager, org.apache.spark.storage.BlockManager, SecurityManager, org.apache.spark.metrics.MetricsSystem, MemoryManager, org.apache.spark.scheduler.OutputCommitCoordinator, SparkConf) - Constructor for class org.apache.spark.SparkEnv
-
- sparkError(String) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- SparkErrorUtils - Interface in org.apache.spark.util
-
- sparkEventFromJson(String) - Static method in class org.apache.spark.util.JsonProtocol
-
- sparkEventFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- sparkEventToJsonString(SparkListenerEvent) - Static method in class org.apache.spark.util.JsonProtocol
-
------------------------------------------------- *
JSON serialization methods for SparkListenerEvents |
- SparkException - Exception in org.apache.spark
-
- SparkException(String, Throwable, Option<String>, Map<String, String>, QueryContext[]) - Constructor for exception org.apache.spark.SparkException
-
- SparkException(String, Throwable) - Constructor for exception org.apache.spark.SparkException
-
- SparkException(String) - Constructor for exception org.apache.spark.SparkException
-
- SparkException(String, Map<String, String>, Throwable, QueryContext[], String) - Constructor for exception org.apache.spark.SparkException
-
- SparkException(String, Map<String, String>, Throwable) - Constructor for exception org.apache.spark.SparkException
-
- SparkException(String, Map<String, String>, Throwable, QueryContext[]) - Constructor for exception org.apache.spark.SparkException
-
- SparkExecutorInfo - Interface in org.apache.spark
-
Exposes information about Spark Executors.
- SparkExecutorInfoImpl - Class in org.apache.spark
-
- SparkExecutorInfoImpl(String, int, long, int, long, long, long, long) - Constructor for class org.apache.spark.SparkExecutorInfoImpl
-
- SparkExitCode - Class in org.apache.spark.util
-
- SparkExitCode() - Constructor for class org.apache.spark.util.SparkExitCode
-
- SparkFiles - Class in org.apache.spark
-
Resolves paths to files added through SparkContext.addFile()
.
- SparkFiles() - Constructor for class org.apache.spark.SparkFiles
-
- SparkFileUtils - Interface in org.apache.spark.util
-
- SparkFilterApi - Class in org.apache.parquet.filter2.predicate
-
TODO (PARQUET-1809): This is a temporary workaround; it is intended to be moved to Parquet.
- SparkFilterApi() - Constructor for class org.apache.parquet.filter2.predicate.SparkFilterApi
-
- SparkFirehoseListener - Class in org.apache.spark
-
Class that allows users to receive all SparkListener events.
- SparkFirehoseListener() - Constructor for class org.apache.spark.SparkFirehoseListener
-
- SparkHadoopMapRedUtil - Class in org.apache.spark.mapred
-
- SparkHadoopMapRedUtil() - Constructor for class org.apache.spark.mapred.SparkHadoopMapRedUtil
-
- sparkJavaOpts(SparkConf, Function1<String, Object>) - Static method in class org.apache.spark.util.Utils
-
Convert all spark properties set in the given SparkConf to a sequence of java options.
- SparkJobInfo - Interface in org.apache.spark
-
Exposes information about Spark Jobs.
- SparkJobInfoImpl - Class in org.apache.spark
-
- SparkJobInfoImpl(int, int[], JobExecutionStatus) - Constructor for class org.apache.spark.SparkJobInfoImpl
-
- SparkLauncher - Class in org.apache.spark.launcher
-
Launcher for Spark applications.
- SparkLauncher() - Constructor for class org.apache.spark.launcher.SparkLauncher
-
- SparkLauncher(Map<String, String>) - Constructor for class org.apache.spark.launcher.SparkLauncher
-
Creates a launcher that will set the given environment variables in the child.
- SparkListener - Class in org.apache.spark.scheduler
-
:: DeveloperApi ::
A default implementation for SparkListenerInterface
that has no-op implementations for
all callbacks.
- SparkListener() - Constructor for class org.apache.spark.scheduler.SparkListener
-
- SparkListenerApplicationEnd - Class in org.apache.spark.scheduler
-
- SparkListenerApplicationEnd(long) - Constructor for class org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- SparkListenerApplicationStart - Class in org.apache.spark.scheduler
-
- SparkListenerApplicationStart(String, Option<String>, long, String, Option<String>, Option<Map<String, String>>, Option<Map<String, String>>) - Constructor for class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- SparkListenerBlockManagerAdded - Class in org.apache.spark.scheduler
-
- SparkListenerBlockManagerAdded(long, BlockManagerId, long, Option<Object>, Option<Object>) - Constructor for class org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- SparkListenerBlockManagerRemoved - Class in org.apache.spark.scheduler
-
- SparkListenerBlockManagerRemoved(long, BlockManagerId) - Constructor for class org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- SparkListenerBlockUpdated - Class in org.apache.spark.scheduler
-
- SparkListenerBlockUpdated(BlockUpdatedInfo) - Constructor for class org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- SparkListenerBus - Interface in org.apache.spark.scheduler
-
- SparkListenerEnvironmentUpdate - Class in org.apache.spark.scheduler
-
- SparkListenerEnvironmentUpdate(Map<String, Seq<Tuple2<String, String>>>) - Constructor for class org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- SparkListenerEvent - Interface in org.apache.spark.scheduler
-
- SparkListenerExecutorAdded - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorAdded(long, String, ExecutorInfo) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- SparkListenerExecutorBlacklisted - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorBlacklisted(long, String, int) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
Deprecated.
- SparkListenerExecutorBlacklistedForStage - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorBlacklistedForStage(long, String, int, int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
-
Deprecated.
- SparkListenerExecutorExcluded - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorExcluded(long, String, int) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorExcluded
-
- SparkListenerExecutorExcludedForStage - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorExcludedForStage(long, String, int, int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
-
- SparkListenerExecutorMetricsUpdate - Class in org.apache.spark.scheduler
-
Periodic updates from executors.
- SparkListenerExecutorMetricsUpdate(String, Seq<Tuple4<Object, Object, Object, Seq<AccumulableInfo>>>, Map<Tuple2<Object, Object>, ExecutorMetrics>) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- SparkListenerExecutorRemoved - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorRemoved(long, String, String) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- SparkListenerExecutorUnblacklisted - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorUnblacklisted(long, String) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
Deprecated.
- SparkListenerExecutorUnexcluded - Class in org.apache.spark.scheduler
-
- SparkListenerExecutorUnexcluded(long, String) - Constructor for class org.apache.spark.scheduler.SparkListenerExecutorUnexcluded
-
- SparkListenerInterface - Interface in org.apache.spark.scheduler
-
Interface for listening to events from the Spark scheduler.
- SparkListenerJobEnd - Class in org.apache.spark.scheduler
-
- SparkListenerJobEnd(int, long, JobResult) - Constructor for class org.apache.spark.scheduler.SparkListenerJobEnd
-
- SparkListenerJobStart - Class in org.apache.spark.scheduler
-
- SparkListenerJobStart(int, long, Seq<StageInfo>, Properties) - Constructor for class org.apache.spark.scheduler.SparkListenerJobStart
-
- SparkListenerLogStart - Class in org.apache.spark.scheduler
-
An internal class that describes the metadata of an event log.
- SparkListenerLogStart(String) - Constructor for class org.apache.spark.scheduler.SparkListenerLogStart
-
- SparkListenerMiscellaneousProcessAdded - Class in org.apache.spark.scheduler
-
- SparkListenerMiscellaneousProcessAdded(long, String, MiscellaneousProcessDetails) - Constructor for class org.apache.spark.scheduler.SparkListenerMiscellaneousProcessAdded
-
- SparkListenerNodeBlacklisted - Class in org.apache.spark.scheduler
-
- SparkListenerNodeBlacklisted(long, String, int) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
Deprecated.
- SparkListenerNodeBlacklistedForStage - Class in org.apache.spark.scheduler
-
- SparkListenerNodeBlacklistedForStage(long, String, int, int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
-
Deprecated.
- SparkListenerNodeExcluded - Class in org.apache.spark.scheduler
-
- SparkListenerNodeExcluded(long, String, int) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeExcluded
-
- SparkListenerNodeExcludedForStage - Class in org.apache.spark.scheduler
-
- SparkListenerNodeExcludedForStage(long, String, int, int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
-
- SparkListenerNodeUnblacklisted - Class in org.apache.spark.scheduler
-
- SparkListenerNodeUnblacklisted(long, String) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
Deprecated.
- SparkListenerNodeUnexcluded - Class in org.apache.spark.scheduler
-
- SparkListenerNodeUnexcluded(long, String) - Constructor for class org.apache.spark.scheduler.SparkListenerNodeUnexcluded
-
- SparkListenerResourceProfileAdded - Class in org.apache.spark.scheduler
-
- SparkListenerResourceProfileAdded(ResourceProfile) - Constructor for class org.apache.spark.scheduler.SparkListenerResourceProfileAdded
-
- SparkListenerSpeculativeTaskSubmitted - Class in org.apache.spark.scheduler
-
- SparkListenerSpeculativeTaskSubmitted(int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- SparkListenerSpeculativeTaskSubmitted(int, int, int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- SparkListenerStageCompleted - Class in org.apache.spark.scheduler
-
- SparkListenerStageCompleted(StageInfo) - Constructor for class org.apache.spark.scheduler.SparkListenerStageCompleted
-
- SparkListenerStageExecutorMetrics - Class in org.apache.spark.scheduler
-
Peak metric values for the executor for the stage, written to the history log at stage
completion.
- SparkListenerStageExecutorMetrics(String, int, int, ExecutorMetrics) - Constructor for class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
-
- SparkListenerStageSubmitted - Class in org.apache.spark.scheduler
-
- SparkListenerStageSubmitted(StageInfo, Properties) - Constructor for class org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- SparkListenerTaskEnd - Class in org.apache.spark.scheduler
-
- SparkListenerTaskEnd(int, int, String, TaskEndReason, TaskInfo, ExecutorMetrics, TaskMetrics) - Constructor for class org.apache.spark.scheduler.SparkListenerTaskEnd
-
- SparkListenerTaskGettingResult - Class in org.apache.spark.scheduler
-
- SparkListenerTaskGettingResult(TaskInfo) - Constructor for class org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- SparkListenerTaskStart - Class in org.apache.spark.scheduler
-
- SparkListenerTaskStart(int, int, TaskInfo) - Constructor for class org.apache.spark.scheduler.SparkListenerTaskStart
-
- SparkListenerUnpersistRDD - Class in org.apache.spark.scheduler
-
- SparkListenerUnpersistRDD(int) - Constructor for class org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- SparkListenerUnschedulableTaskSetAdded - Class in org.apache.spark.scheduler
-
- SparkListenerUnschedulableTaskSetAdded(int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetAdded
-
- SparkListenerUnschedulableTaskSetRemoved - Class in org.apache.spark.scheduler
-
- SparkListenerUnschedulableTaskSetRemoved(int, int) - Constructor for class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetRemoved
-
- SparkMasterRegex - Class in org.apache.spark
-
A collection of regexes for extracting information from the master string.
- SparkMasterRegex() - Constructor for class org.apache.spark.SparkMasterRegex
-
- SparkPath - Class in org.apache.spark.paths
-
A canonical representation of a file path.
- SparkPath() - Constructor for class org.apache.spark.paths.SparkPath
-
- SparkPlugin - Interface in org.apache.spark.api.plugin
-
:: DeveloperApi ::
A plugin that can be dynamically loaded into a Spark application.
- sparkProperties() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig
-
- sparkProperties() - Method in class org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- sparkRPackagePath(boolean) - Static method in class org.apache.spark.api.r.RUtils
-
Get the list of paths for R packages in various deployment modes, of which the first
path is for the SparkR package itself.
- SparkSchemaUtils - Class in org.apache.spark.util
-
Utils for handling schemas.
- SparkSchemaUtils() - Constructor for class org.apache.spark.util.SparkSchemaUtils
-
- SparkSerDeUtils - Interface in org.apache.spark.util
-
- sparkSession() - Method in interface org.apache.spark.ml.util.BaseReadWrite
-
Returns the user-specified Spark Session or the default.
- sparkSession() - Method in class org.apache.spark.sql.Dataset
-
- SparkSession - Class in org.apache.spark.sql
-
The entry point to programming Spark with the Dataset and DataFrame API.
- sparkSession() - Method in class org.apache.spark.sql.SQLContext
-
- sparkSession() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
-
Returns the SparkSession
associated with this
.
- SparkSession.Builder - Class in org.apache.spark.sql
-
- SparkSession.implicits$ - Class in org.apache.spark.sql
-
(Scala-specific) Implicit methods available in Scala for converting
common Scala objects into DataFrame
s.
- SparkSessionExtensions - Class in org.apache.spark.sql
-
:: Experimental ::
Holder for injection points to the
SparkSession
.
- SparkSessionExtensions() - Constructor for class org.apache.spark.sql.SparkSessionExtensions
-
- SparkSessionExtensionsProvider - Interface in org.apache.spark.sql
-
:: Unstable ::
- SparkShutdownHook - Class in org.apache.spark.util
-
- SparkShutdownHook(int, Function0<BoxedUnit>) - Constructor for class org.apache.spark.util.SparkShutdownHook
-
- SparkStageInfo - Interface in org.apache.spark
-
Exposes information about Spark Stages.
- SparkStageInfoImpl - Class in org.apache.spark
-
- SparkStageInfoImpl(int, int, long, String, int, int, int, int) - Constructor for class org.apache.spark.SparkStageInfoImpl
-
- SparkStatusTracker - Class in org.apache.spark
-
Low-level status reporting APIs for monitoring job and stage progress.
- SparkThreadUtils - Class in org.apache.spark.util
-
- SparkThreadUtils() - Constructor for class org.apache.spark.util.SparkThreadUtils
-
- SparkThrowable - Interface in org.apache.spark
-
Interface mixed into Throwables thrown from Spark.
- SparkThrowableHelper - Class in org.apache.spark
-
Companion object used by instances of
SparkThrowable
to access error class information and
construct error messages.
- SparkThrowableHelper() - Constructor for class org.apache.spark.SparkThrowableHelper
-
- sparkUpgradeInReadingDatesError(String, String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- sparkUpgradeInWritingDatesError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- sparkUser() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- sparkUser() - Method in class org.apache.spark.scheduler.SparkListenerApplicationStart
-
- sparkUser() - Method in class org.apache.spark.SparkContext
-
- sparkUser() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- sparkVersion() - Method in class org.apache.spark.scheduler.SparkListenerLogStart
-
- sparse(int, int, int[], int[], double[]) - Static method in class org.apache.spark.ml.linalg.Matrices
-
Creates a column-major sparse matrix in Compressed Sparse Column (CSC) format.
- sparse(int, int[], double[]) - Static method in class org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector providing its index array and value array.
- sparse(int, Seq<Tuple2<Object, Object>>) - Static method in class org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs.
- sparse(int, Iterable<Tuple2<Integer, Double>>) - Static method in class org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs in a Java friendly way.
- sparse(int, int, int[], int[], double[]) - Static method in class org.apache.spark.mllib.linalg.Matrices
-
Creates a column-major sparse matrix in Compressed Sparse Column (CSC) format.
- sparse(int, int[], double[]) - Static method in class org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector providing its index array and value array.
- sparse(int, Seq<Tuple2<Object, Object>>) - Static method in class org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs.
- sparse(int, Iterable<Tuple2<Integer, Double>>) - Static method in class org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs in a Java friendly way.
- SparseMatrix - Class in org.apache.spark.ml.linalg
-
Column-major sparse matrix.
- SparseMatrix(int, int, int[], int[], double[], boolean) - Constructor for class org.apache.spark.ml.linalg.SparseMatrix
-
- SparseMatrix(int, int, int[], int[], double[]) - Constructor for class org.apache.spark.ml.linalg.SparseMatrix
-
Column-major sparse matrix.
- SparseMatrix - Class in org.apache.spark.mllib.linalg
-
Column-major sparse matrix.
- SparseMatrix(int, int, int[], int[], double[], boolean) - Constructor for class org.apache.spark.mllib.linalg.SparseMatrix
-
- SparseMatrix(int, int, int[], int[], double[]) - Constructor for class org.apache.spark.mllib.linalg.SparseMatrix
-
Column-major sparse matrix.
- SparseVector - Class in org.apache.spark.ml.linalg
-
A sparse vector represented by an index array and a value array.
- SparseVector(int, int[], double[]) - Constructor for class org.apache.spark.ml.linalg.SparseVector
-
- SparseVector - Class in org.apache.spark.mllib.linalg
-
A sparse vector represented by an index array and a value array.
- SparseVector(int, int[], double[]) - Constructor for class org.apache.spark.mllib.linalg.SparseVector
-
- SPARSITY() - Static method in class org.apache.spark.ml.attribute.AttributeKeys
-
- sparsity() - Method in class org.apache.spark.ml.attribute.NumericAttribute
-
- sparsity() - Method in interface org.apache.spark.mllib.linalg.Vector
-
Returns the ratio of number of zeros by total number of values.
- spdiag(Vector) - Static method in class org.apache.spark.ml.linalg.SparseMatrix
-
Generate a diagonal matrix in SparseMatrix
format from the supplied values.
- spdiag(Vector) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a diagonal matrix in SparseMatrix
format from the supplied values.
- SpearmanCorrelation - Class in org.apache.spark.mllib.stat.correlation
-
Compute Spearman's correlation for two RDDs of the type RDD[Double] or the correlation matrix
for an RDD of the type RDD[Vector].
- SpearmanCorrelation() - Constructor for class org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
- SpecialLengths - Class in org.apache.spark.api.r
-
- SpecialLengths() - Constructor for class org.apache.spark.api.r.SpecialLengths
-
- specifyingDBInCreateTempFuncError(String, SqlBaseParser.CreateFunctionContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- specifyPartitionNotAllowedWhenTableSchemaNotDefinedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- SPECULATION_SUMMARY_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- SpeculationStageSummary - Class in org.apache.spark.status.api.v1
-
- speculationStageSummary() - Method in class org.apache.spark.status.LiveStage
-
- speculationSummary() - Method in class org.apache.spark.status.api.v1.StageData
-
- speculative() - Method in class org.apache.spark.scheduler.TaskInfo
-
- speculative() - Method in class org.apache.spark.status.api.v1.TaskData
-
- SPECULATIVE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- SPECULATIVE_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- speye(int) - Static method in class org.apache.spark.ml.linalg.Matrices
-
Generate a sparse Identity Matrix in Matrix
format.
- speye(int) - Static method in class org.apache.spark.ml.linalg.SparseMatrix
-
Generate an Identity Matrix in SparseMatrix
format.
- speye(int) - Static method in class org.apache.spark.mllib.linalg.Matrices
-
Generate a sparse Identity Matrix in Matrix
format.
- speye(int) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
-
Generate an Identity Matrix in SparseMatrix
format.
- SpillListener - Class in org.apache.spark
-
A SparkListener
that detects whether spills have occurred in Spark jobs.
- SpillListener() - Constructor for class org.apache.spark.SpillListener
-
- split() - Method in class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- split() - Method in class org.apache.spark.ml.tree.InternalNode
-
- Split - Interface in org.apache.spark.ml.tree
-
Interface for a "Split," which specifies a test made at a decision tree node
to choose the left or right path.
- split() - Method in class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- split() - Method in class org.apache.spark.mllib.tree.model.Node
-
- Split - Class in org.apache.spark.mllib.tree.model
-
Split applied to a feature
param: feature feature index
param: threshold Threshold for continuous feature.
- Split(int, double, Enumeration.Value, List<Object>) - Constructor for class org.apache.spark.mllib.tree.model.Split
-
- split(Column, String) - Static method in class org.apache.spark.sql.functions
-
Splits str around matches of the given pattern.
- split(Column, String, int) - Static method in class org.apache.spark.sql.functions
-
Splits str around matches of the given pattern.
- split_part(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Splits str
by delimiter and return requested part of the split (1-based).
- splitAndCountPartitions(Iterator<String>) - Static method in class org.apache.spark.streaming.util.RawTextHelper
-
Splits lines and counts the words.
- splitCommandString(String) - Static method in class org.apache.spark.util.Utils
-
Split a string of potentially quoted arguments from the command line the way that a shell
would do it to determine arguments to a command.
- SplitData(int, double[], int) - Constructor for class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- SplitData(int, double, int, Seq<Object>) - Constructor for class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- SplitData$() - Constructor for class org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
-
- SplitData$() - Constructor for class org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData$
-
- splitIndex() - Method in class org.apache.spark.storage.RDDBlockId
-
- SplitInfo - Class in org.apache.spark.scheduler
-
- SplitInfo(Class<?>, String, String, long, Object) - Constructor for class org.apache.spark.scheduler.SplitInfo
-
- splits() - Method in class org.apache.spark.ml.feature.Bucketizer
-
Parameter for mapping continuous features into buckets.
- splitsArray() - Method in class org.apache.spark.ml.feature.Bucketizer
-
Parameter for specifying multiple splits parameters.
- spr(double, Vector, DenseVector) - Static method in class org.apache.spark.ml.linalg.BLAS
-
Adds alpha * x * x.t to a matrix in-place.
- spr(double, Vector, double[]) - Static method in class org.apache.spark.ml.linalg.BLAS
-
Adds alpha * v * v.t to a matrix in-place.
- spr(double, Vector, DenseVector) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
Adds alpha * v * v.t to a matrix in-place.
- spr(double, Vector, double[]) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
Adds alpha * v * v.t to a matrix in-place.
- sprand(int, int, double, Random) - Static method in class org.apache.spark.ml.linalg.Matrices
-
Generate a SparseMatrix
consisting of i.i.d.
uniform random numbers.
- sprand(int, int, double, Random) - Static method in class org.apache.spark.ml.linalg.SparseMatrix
-
Generate a SparseMatrix
consisting of i.i.d
.
- sprand(int, int, double, Random) - Static method in class org.apache.spark.mllib.linalg.Matrices
-
Generate a SparseMatrix
consisting of i.i.d.
uniform random numbers.
- sprand(int, int, double, Random) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a SparseMatrix
consisting of i.i.d
.
- sprandn(int, int, double, Random) - Static method in class org.apache.spark.ml.linalg.Matrices
-
Generate a SparseMatrix
consisting of i.i.d.
gaussian random numbers.
- sprandn(int, int, double, Random) - Static method in class org.apache.spark.ml.linalg.SparseMatrix
-
Generate a SparseMatrix
consisting of i.i.d
.
- sprandn(int, int, double, Random) - Static method in class org.apache.spark.mllib.linalg.Matrices
-
Generate a SparseMatrix
consisting of i.i.d.
gaussian random numbers.
- sprandn(int, int, double, Random) - Static method in class org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a SparseMatrix
consisting of i.i.d
.
- sqdist(Vector, Vector) - Static method in class org.apache.spark.ml.linalg.Vectors
-
Returns the squared distance between two Vectors.
- sqdist(Vector, Vector) - Static method in class org.apache.spark.mllib.linalg.Vectors
-
Returns the squared distance between two Vectors.
- sql(String, Object) - Method in class org.apache.spark.sql.SparkSession
-
Executes a SQL query substituting positional parameters by the given arguments,
returning the result as a DataFrame
.
- sql(String, Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession
-
Executes a SQL query substituting named parameters by the given arguments,
returning the result as a DataFrame
.
- sql(String, Map<String, Object>) - Method in class org.apache.spark.sql.SparkSession
-
Executes a SQL query substituting named parameters by the given arguments,
returning the result as a DataFrame
.
- sql(String) - Method in class org.apache.spark.sql.SparkSession
-
Executes a SQL query using Spark, returning the result as a DataFrame
.
- sql(String) - Method in class org.apache.spark.sql.SQLContext
-
- sql() - Method in class org.apache.spark.sql.types.ArrayType
-
- sql() - Static method in class org.apache.spark.sql.types.BinaryType
-
- sql() - Static method in class org.apache.spark.sql.types.BooleanType
-
- sql() - Static method in class org.apache.spark.sql.types.ByteType
-
- sql() - Static method in class org.apache.spark.sql.types.CalendarIntervalType
-
- sql() - Method in class org.apache.spark.sql.types.DataType
-
- sql() - Static method in class org.apache.spark.sql.types.DateType
-
- sql() - Method in class org.apache.spark.sql.types.DecimalType
-
- sql() - Static method in class org.apache.spark.sql.types.DoubleType
-
- sql() - Static method in class org.apache.spark.sql.types.FloatType
-
- sql() - Static method in class org.apache.spark.sql.types.IntegerType
-
- sql() - Static method in class org.apache.spark.sql.types.LongType
-
- sql() - Method in class org.apache.spark.sql.types.MapType
-
- sql() - Static method in class org.apache.spark.sql.types.NullType
-
- sql() - Static method in class org.apache.spark.sql.types.ShortType
-
- sql() - Static method in class org.apache.spark.sql.types.StringType
-
- sql() - Method in class org.apache.spark.sql.types.StructType
-
- sql() - Static method in class org.apache.spark.sql.types.TimestampNTZType
-
- sql() - Static method in class org.apache.spark.sql.types.TimestampType
-
- sql() - Method in class org.apache.spark.sql.types.UserDefinedType
-
- SQL_EXECUTION_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobDataWrapper
-
- sqlConfigNotFoundError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- sqlContext() - Method in interface org.apache.spark.ml.util.BaseReadWrite
-
Returns the user-specified SQL context or the default.
- sqlContext() - Method in class org.apache.spark.sql.Dataset
-
- sqlContext() - Method in class org.apache.spark.sql.sources.BaseRelation
-
- sqlContext() - Method in class org.apache.spark.sql.SparkSession
-
A wrapped version of this session in the form of a
SQLContext
, for backward compatibility.
- SQLContext - Class in org.apache.spark.sql
-
The entry point for working with structured data (rows and columns) in Spark 1.x.
- SQLContext(SparkContext) - Constructor for class org.apache.spark.sql.SQLContext
-
- SQLContext(JavaSparkContext) - Constructor for class org.apache.spark.sql.SQLContext
-
- SQLContext.implicits$ - Class in org.apache.spark.sql
-
(Scala-specific) Implicit methods available in Scala for converting
common Scala objects into DataFrame
s.
- SQLDataTypes - Class in org.apache.spark.ml.linalg
-
SQL data types for vectors and matrices.
- SQLDataTypes() - Constructor for class org.apache.spark.ml.linalg.SQLDataTypes
-
- SQLImplicits - Class in org.apache.spark.sql
-
A collection of implicit methods for converting common Scala objects into
Dataset
s.
- SQLImplicits() - Constructor for class org.apache.spark.sql.SQLImplicits
-
- SQLImplicits.StringToColumn - Class in org.apache.spark.sql
-
Converts $"col name" into a
Column
.
- SQLOpenHashSet<T> - Class in org.apache.spark.sql.util
-
- SQLOpenHashSet(int, double, ClassTag<T>) - Constructor for class org.apache.spark.sql.util.SQLOpenHashSet
-
- SQLOpenHashSet(int, ClassTag<T>) - Constructor for class org.apache.spark.sql.util.SQLOpenHashSet
-
- SQLOpenHashSet(ClassTag<T>) - Constructor for class org.apache.spark.sql.util.SQLOpenHashSet
-
- SQLPlanMetricSerializer - Class in org.apache.spark.status.protobuf.sql
-
- SQLPlanMetricSerializer() - Constructor for class org.apache.spark.status.protobuf.sql.SQLPlanMetricSerializer
-
- sqlState() - Method in class org.apache.spark.ErrorInfo
-
- sqlStatementUnsupportedError(String, Origin) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- SQLTransformer - Class in org.apache.spark.ml.feature
-
Implements the transformations which are defined by SQL statement.
- SQLTransformer(String) - Constructor for class org.apache.spark.ml.feature.SQLTransformer
-
- SQLTransformer() - Constructor for class org.apache.spark.ml.feature.SQLTransformer
-
- sqlType() - Method in class org.apache.spark.mllib.linalg.VectorUDT
-
- sqlType() - Method in class org.apache.spark.sql.types.UserDefinedType
-
Underlying storage type for this UDT
- SQLUserDefinedType - Annotation Type in org.apache.spark.sql.types
-
::DeveloperApi::
A user-defined type which can be automatically recognized by a SQLContext and registered.
- SQLUtils - Class in org.apache.spark.sql.api.r
-
- SQLUtils() - Constructor for class org.apache.spark.sql.api.r.SQLUtils
-
- sqrt(Column) - Static method in class org.apache.spark.sql.functions
-
Computes the square root of the specified float value.
- sqrt(String) - Static method in class org.apache.spark.sql.functions
-
Computes the square root of the specified float value.
- Sqrt$() - Constructor for class org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
-
- SquaredError - Class in org.apache.spark.mllib.tree.loss
-
Class for squared error loss calculation.
- SquaredError() - Constructor for class org.apache.spark.mllib.tree.loss.SquaredError
-
- SquaredEuclideanSilhouette - Class in org.apache.spark.ml.evaluation
-
SquaredEuclideanSilhouette computes the average of the
Silhouette over all the data of the dataset, which is
a measure of how appropriately the data have been clustered.
- SquaredEuclideanSilhouette() - Constructor for class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
- SquaredEuclideanSilhouette.ClusterStats - Class in org.apache.spark.ml.evaluation
-
- SquaredEuclideanSilhouette.ClusterStats$ - Class in org.apache.spark.ml.evaluation
-
- SquaredL2Updater - Class in org.apache.spark.mllib.optimization
-
Updater for L2 regularized problems.
- SquaredL2Updater() - Constructor for class org.apache.spark.mllib.optimization.SquaredL2Updater
-
- squaredNormSum() - Method in class org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
-
- Src - Static variable in class org.apache.spark.graphx.TripletFields
-
Expose the source and edge fields but not the destination field.
- srcAttr() - Method in class org.apache.spark.graphx.EdgeContext
-
The vertex attribute of the edge's source vertex.
- srcAttr() - Method in class org.apache.spark.graphx.EdgeTriplet
-
The source vertex attribute
- srcAttr() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- srcCol() - Method in class org.apache.spark.ml.clustering.PowerIterationClustering
-
- srcCol() - Method in interface org.apache.spark.ml.clustering.PowerIterationClusteringParams
-
Param for the name of the input column for source vertex IDs.
- srcId() - Method in class org.apache.spark.graphx.Edge
-
- srcId() - Method in class org.apache.spark.graphx.EdgeContext
-
The vertex id of the edge's source vertex.
- srcId() - Method in class org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- srdd() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
- ssc() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
- stack(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Separates col1
, ..., colk
into n
rows.
- stackTrace() - Method in class org.apache.spark.ExceptionFailure
-
- StackTrace - Class in org.apache.spark.status.api.v1
-
- StackTrace(Seq<String>) - Constructor for class org.apache.spark.status.api.v1.StackTrace
-
- stackTrace() - Method in class org.apache.spark.status.api.v1.ThreadStackTrace
-
- stackTraceFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- stackTraceToJson(StackTraceElement[], JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- stackTraceToString(Throwable) - Method in interface org.apache.spark.util.SparkErrorUtils
-
- stackTraceToString(Throwable) - Static method in class org.apache.spark.util.Utils
-
- stage() - Method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- STAGE() - Static method in class org.apache.spark.status.TaskIndexNames
-
- STAGE_ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- STAGE_ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- STAGE_ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- STAGE_ATTEMPT_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- STAGE_DAG() - Static method in class org.apache.spark.ui.ToolTips
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.CachedQuantile
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummaryWrapper
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDOperationGraphWrapper
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SpeculationStageSummaryWrapper
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- STAGE_ID_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- STAGE_IDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- STAGE_IDS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.PoolData
-
- STAGE_STATUS_ACTIVE_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_ACTIVE = 1;
- STAGE_STATUS_COMPLETE_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_COMPLETE = 2;
- STAGE_STATUS_FAILED_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_FAILED = 3;
- STAGE_STATUS_PENDING_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_PENDING = 4;
- STAGE_STATUS_SKIPPED_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_SKIPPED = 5;
- STAGE_STATUS_UNSPECIFIED_VALUE - Static variable in enum org.apache.spark.status.protobuf.StoreTypes.StageStatus
-
STAGE_STATUS_UNSPECIFIED = 0;
- STAGE_TIMELINE() - Static method in class org.apache.spark.ui.ToolTips
-
- stageAttempt() - Method in class org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- stageAttemptId() - Method in class org.apache.spark.ContextBarrierId
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
-
Deprecated.
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
-
Deprecated.
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerTaskEnd
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerTaskStart
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetAdded
-
- stageAttemptId() - Method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetRemoved
-
- stageAttemptNumber() - Method in class org.apache.spark.BarrierTaskContext
-
- stageAttemptNumber() - Method in class org.apache.spark.TaskContext
-
How many times the stage that this task belongs to has been attempted.
- stageCompletedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- stageCompletedToJson(SparkListenerStageCompleted, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- stageCreate(Identifier, StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
Deprecated.
- stageCreate(Identifier, Column[], Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
Stage the creation of a table, preparing it to be committed into the metastore.
- stageCreateOrReplace(Identifier, StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
- stageCreateOrReplace(Identifier, Column[], Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
- StageData - Class in org.apache.spark.status.api.v1
-
- StagedTable - Interface in org.apache.spark.sql.connector.catalog
-
Represents a table which is staged for being committed to the metastore.
- stageExecutorMetricsFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- stageExecutorMetricsToJson(SparkListenerStageExecutorMetrics, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- stageFailed(String) - Method in class org.apache.spark.scheduler.StageInfo
-
- stageId() - Method in class org.apache.spark.BarrierTaskContext
-
- stageId() - Method in class org.apache.spark.ContextBarrierId
-
- stageId() - Method in interface org.apache.spark.scheduler.Schedulable
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorBlacklistedForStage
-
Deprecated.
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerExecutorExcludedForStage
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerNodeBlacklistedForStage
-
Deprecated.
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerNodeExcludedForStage
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerStageExecutorMetrics
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerTaskEnd
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerTaskStart
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetAdded
-
- stageId() - Method in class org.apache.spark.scheduler.SparkListenerUnschedulableTaskSetRemoved
-
- stageId() - Method in class org.apache.spark.scheduler.StageInfo
-
- stageId() - Method in interface org.apache.spark.SparkStageInfo
-
- stageId() - Method in class org.apache.spark.SparkStageInfoImpl
-
- stageId() - Method in class org.apache.spark.status.api.v1.StageData
-
- stageId() - Method in class org.apache.spark.TaskContext
-
The ID of the stage that this task belong to.
- stageIds() - Method in class org.apache.spark.scheduler.SparkListenerJobStart
-
- stageIds() - Method in interface org.apache.spark.SparkJobInfo
-
- stageIds() - Method in class org.apache.spark.SparkJobInfoImpl
-
- stageIds() - Method in class org.apache.spark.status.api.v1.JobData
-
- stageIds() - Method in class org.apache.spark.status.LiveJob
-
- stageIds() - Method in class org.apache.spark.status.SchedulerPool
-
- stageInfo() - Method in class org.apache.spark.scheduler.SparkListenerStageCompleted
-
- stageInfo() - Method in class org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- StageInfo - Class in org.apache.spark.scheduler
-
:: DeveloperApi ::
Stores information about a stage to pass from the scheduler to SparkListeners.
- StageInfo(int, int, String, int, Seq<RDDInfo>, Seq<Object>, String, TaskMetrics, Seq<Seq<TaskLocation>>, Option<Object>, int, boolean, int) - Constructor for class org.apache.spark.scheduler.StageInfo
-
- stageInfoFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
--------------------------------------------------------------------- *
JSON deserialization methods for classes SparkListenerEvents depend on |
- stageInfos() - Method in class org.apache.spark.scheduler.SparkListenerJobStart
-
- stageInfoToJson(StageInfo, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
------------------------------------------------------------------- *
JSON serialization methods for classes SparkListenerEvents depend on |
- stageName() - Method in class org.apache.spark.ml.clustering.InternalKMeansModelWriter
-
- stageName() - Method in class org.apache.spark.ml.clustering.PMMLKMeansModelWriter
-
- stageName() - Method in class org.apache.spark.ml.regression.InternalLinearRegressionModelWriter
-
- stageName() - Method in class org.apache.spark.ml.regression.PMMLLinearRegressionModelWriter
-
- stageName() - Method in interface org.apache.spark.ml.util.MLFormatRegister
-
The string that represents the stage type that this writer supports.
- stageReplace(Identifier, StructType, Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
- stageReplace(Identifier, Column[], Transform[], Map<String, String>) - Method in interface org.apache.spark.sql.connector.catalog.StagingTableCatalog
-
- stages() - Method in class org.apache.spark.ml.Pipeline
-
param for pipeline stages
- stages() - Method in class org.apache.spark.ml.PipelineModel
-
- STAGES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- StageStatus - Enum in org.apache.spark.status.api.v1
-
- StageStatusSerializer - Class in org.apache.spark.status.protobuf
-
- StageStatusSerializer() - Constructor for class org.apache.spark.status.protobuf.StageStatusSerializer
-
- stageSubmittedFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- stageSubmittedToJson(SparkListenerStageSubmitted, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- StagingTableCatalog - Interface in org.apache.spark.sql.connector.catalog
-
An optional mix-in for implementations of
TableCatalog
that support staging creation of
the a table before committing the table's metadata along with its contents in CREATE TABLE AS
SELECT or REPLACE TABLE AS SELECT operations.
- STANDARD() - Static method in class org.apache.spark.ErrorMessageFormat
-
- standardization() - Method in class org.apache.spark.ml.classification.LinearSVC
-
- standardization() - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
- standardization() - Method in class org.apache.spark.ml.classification.LogisticRegression
-
- standardization() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
- standardization() - Method in interface org.apache.spark.ml.param.shared.HasStandardization
-
Param for whether to standardize the training features before fitting the model.
- standardization() - Method in class org.apache.spark.ml.regression.LinearRegression
-
- standardization() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
- StandardNormalGenerator - Class in org.apache.spark.mllib.random
-
Generates i.i.d.
- StandardNormalGenerator() - Constructor for class org.apache.spark.mllib.random.StandardNormalGenerator
-
- StandardScaler - Class in org.apache.spark.ml.feature
-
Standardizes features by removing the mean and scaling to unit variance using column summary
statistics on the samples in the training set.
- StandardScaler(String) - Constructor for class org.apache.spark.ml.feature.StandardScaler
-
- StandardScaler() - Constructor for class org.apache.spark.ml.feature.StandardScaler
-
- StandardScaler - Class in org.apache.spark.mllib.feature
-
Standardizes features by removing the mean and scaling to unit std using column summary
statistics on the samples in the training set.
- StandardScaler(boolean, boolean) - Constructor for class org.apache.spark.mllib.feature.StandardScaler
-
- StandardScaler() - Constructor for class org.apache.spark.mllib.feature.StandardScaler
-
- StandardScalerModel - Class in org.apache.spark.ml.feature
-
- StandardScalerModel - Class in org.apache.spark.mllib.feature
-
Represents a StandardScaler model that can transform vectors.
- StandardScalerModel(Vector, Vector, boolean, boolean) - Constructor for class org.apache.spark.mllib.feature.StandardScalerModel
-
- StandardScalerModel(Vector, Vector) - Constructor for class org.apache.spark.mllib.feature.StandardScalerModel
-
- StandardScalerModel(Vector) - Constructor for class org.apache.spark.mllib.feature.StandardScalerModel
-
- StandardScalerParams - Interface in org.apache.spark.ml.feature
-
- starExpandDataTypeNotSupportedError(Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- starGraph(SparkContext, int) - Static method in class org.apache.spark.graphx.util.GraphGenerators
-
Create a star graph with vertex 0 being the center.
- starNotAllowedWhenGroupByOrdinalPositionUsedError() - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- start() - Method in interface org.apache.spark.metrics.sink.Sink
-
- start() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
- start() - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- start(String) - Method in class org.apache.spark.sql.streaming.DataStreamWriter
-
Starts the execution of the streaming query, which will continually output results to the given
path as new data arrives.
- start() - Method in class org.apache.spark.sql.streaming.DataStreamWriter
-
Starts the execution of the streaming query, which will continually output results to the given
path as new data arrives.
- start() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Start the execution of the streams.
- start() - Method in class org.apache.spark.streaming.dstream.ConstantInputDStream
-
- start() - Method in class org.apache.spark.streaming.dstream.InputDStream
-
Method called to start receiving data.
- start() - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- start() - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Start the execution of the streams.
- START_OFFSET_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SourceProgress
-
- START_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationAttemptInfo
-
- START_TIMESTAMP_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryData
-
- startApplication(SparkAppHandle.Listener...) - Method in class org.apache.spark.launcher.AbstractLauncher
-
Starts a Spark application.
- startApplication(SparkAppHandle.Listener...) - Method in class org.apache.spark.launcher.InProcessLauncher
-
Starts a Spark application.
- startApplication(SparkAppHandle.Listener...) - Method in class org.apache.spark.launcher.SparkLauncher
-
Starts a Spark application.
- startField() - Method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- startField() - Method in class org.apache.spark.sql.types.YearMonthIntervalType
-
- startIndex() - Method in interface org.apache.spark.QueryContext
-
- startIndexInLevel(int) - Static method in class org.apache.spark.mllib.tree.model.Node
-
Return the index of the first node in the given level.
- startJettyServer(String, int, SSLOptions, SparkConf, String, int) - Static method in class org.apache.spark.ui.JettyUtils
-
Attempt to start a Jetty server bound to the supplied hostName:port using the given
context handlers.
- startOffset() - Method in class org.apache.spark.sql.streaming.SourceProgress
-
- startOffset() - Method in exception org.apache.spark.sql.streaming.StreamingQueryException
-
- startPosition() - Method in exception org.apache.spark.sql.AnalysisException
-
- startReduceId() - Method in class org.apache.spark.storage.ShuffleBlockBatchId
-
- startServiceOnPort(int, Function1<Object, Tuple2<T, Object>>, SparkConf, String) - Static method in class org.apache.spark.util.Utils
-
Attempt to start a service on the given port, or fail after a number of attempts.
- startsWith(Column) - Method in class org.apache.spark.sql.Column
-
String starts with.
- startsWith(String) - Method in class org.apache.spark.sql.Column
-
String starts with another string literal.
- startswith(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns a boolean.
- startTime() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- startTime() - Method in class org.apache.spark.SparkContext
-
- startTime() - Method in class org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- startTime() - Method in class org.apache.spark.status.api.v1.streaming.OutputOperationInfo
-
- startTime() - Method in class org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- startTime() - Method in class org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- stat() - Method in class org.apache.spark.sql.Dataset
-
- StatCounter - Class in org.apache.spark.util
-
A class for tracking the statistics of a set of numbers (count, mean and variance) in a
numerically robust way.
- StatCounter(TraversableOnce<Object>) - Constructor for class org.apache.spark.util.StatCounter
-
- StatCounter() - Constructor for class org.apache.spark.util.StatCounter
-
Initialize the StatCounter with no values.
- state() - Method in class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- state() - Method in class org.apache.spark.scheduler.local.StatusUpdate
-
- State<S> - Class in org.apache.spark.streaming
-
:: Experimental ::
Abstract class for getting and updating the state in mapping function used in the
mapWithState
operation of a
pair DStream
(Scala)
or a
JavaPairDStream
(Java).
- State() - Constructor for class org.apache.spark.streaming.State
-
- STATE_OPERATORS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress
-
- stateChanged(SparkAppHandle) - Method in interface org.apache.spark.launcher.SparkAppHandle.Listener
-
Callback for changes in the handle's state.
- statement() - Method in class org.apache.spark.ml.feature.SQLTransformer
-
SQL statement parameter.
- stateNotDefinedOrAlreadyRemovedError() - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- StateOperatorProgress - Class in org.apache.spark.sql.streaming
-
Information about updates made to stateful operators in a
StreamingQuery
during a trigger.
- StateOperatorProgressSerializer - Class in org.apache.spark.status.protobuf.sql
-
- StateOperatorProgressSerializer() - Constructor for class org.apache.spark.status.protobuf.sql.StateOperatorProgressSerializer
-
- stateOperators() - Method in class org.apache.spark.sql.streaming.StreamingQueryProgress
-
- stateSnapshots() - Method in class org.apache.spark.streaming.api.java.JavaMapWithStateDStream
-
- stateSnapshots() - Method in class org.apache.spark.streaming.dstream.MapWithStateDStream
-
Return a pair DStream where each RDD is the snapshot of the state of all the keys.
- StateSpec<KeyType,ValueType,StateType,MappedType> - Class in org.apache.spark.streaming
-
:: Experimental ::
Abstract class representing all the specifications of the DStream transformation
mapWithState
operation of a
pair DStream
(Scala) or a
JavaPairDStream
(Java).
- StateSpec() - Constructor for class org.apache.spark.streaming.StateSpec
-
- staticPageRank(int, double) - Method in class org.apache.spark.graphx.GraphOps
-
Run PageRank for a fixed number of iterations returning a graph with vertex attributes
containing the PageRank and edge attributes the normalized edge weight.
- staticPageRank(int, double, Graph<Object, Object>) - Method in class org.apache.spark.graphx.GraphOps
-
Run PageRank for a fixed number of iterations returning a graph with vertex attributes
containing the PageRank and edge attributes the normalized edge weight, optionally including
including a previous pageRank computation to be used as a start point for the new iterations
- staticParallelPersonalizedPageRank(long[], int, double) - Method in class org.apache.spark.graphx.GraphOps
-
Run parallel personalized PageRank for a given array of source vertices, such
that all random walks are started relative to the source vertices
- staticPartitionInUserSpecifiedColumnsError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- staticPersonalizedPageRank(long, int, double) - Method in class org.apache.spark.graphx.GraphOps
-
Run Personalized PageRank for a fixed number of iterations with
with all iterations originating at the source node
returning a graph with vertex attributes
containing the PageRank and edge attributes the normalized edge weight.
- StaticSources - Class in org.apache.spark.metrics.source
-
- StaticSources() - Constructor for class org.apache.spark.metrics.source.StaticSources
-
- statistic() - Method in class org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- statistic() - Method in class org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
-
- statistic() - Method in interface org.apache.spark.mllib.stat.test.TestResult
-
Test statistic.
- statisticNotRecognizedError(String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- Statistics - Class in org.apache.spark.mllib.stat
-
API for statistical functions in MLlib.
- Statistics() - Constructor for class org.apache.spark.mllib.stat.Statistics
-
- Statistics - Interface in org.apache.spark.sql.connector.read
-
- stats() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return a
StatCounter
object that captures the mean, variance and
count of the RDD's elements in one operation.
- stats() - Method in class org.apache.spark.mllib.tree.model.Node
-
- stats() - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Return a
StatCounter
object that captures the mean, variance and
count of the RDD's elements in one operation.
- stats() - Method in interface org.apache.spark.sql.columnar.SimpleMetricsCachedBatch
-
Holds stats for each cached column.
- StatsdMetricType - Class in org.apache.spark.metrics.sink
-
- StatsdMetricType() - Constructor for class org.apache.spark.metrics.sink.StatsdMetricType
-
- StatsReportListener - Class in org.apache.spark.scheduler
-
:: DeveloperApi ::
Simple SparkListener that logs a few summary statistics when each stage completes.
- StatsReportListener() - Constructor for class org.apache.spark.scheduler.StatsReportListener
-
- StatsReportListener - Class in org.apache.spark.streaming.scheduler
-
:: DeveloperApi ::
A simple StreamingListener that logs summary statistics across Spark Streaming batches
param: numBatchInfos Number of last batches to consider for generating statistics (default: 10)
- StatsReportListener(int) - Constructor for class org.apache.spark.streaming.scheduler.StatsReportListener
-
- status() - Method in class org.apache.spark.scheduler.TaskInfo
-
- status() - Method in interface org.apache.spark.SparkJobInfo
-
- status() - Method in class org.apache.spark.SparkJobInfoImpl
-
- status() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
-
Returns the current status of the query.
- status() - Method in class org.apache.spark.status.api.v1.JobData
-
- status() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
-
- status() - Method in class org.apache.spark.status.api.v1.StageData
-
- status() - Method in class org.apache.spark.status.api.v1.streaming.BatchInfo
-
- status() - Method in class org.apache.spark.status.api.v1.TaskData
-
- status() - Method in class org.apache.spark.status.LiveJob
-
- status() - Method in class org.apache.spark.status.LiveStage
-
- STATUS() - Static method in class org.apache.spark.status.TaskIndexNames
-
- status() - Method in class org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus
-
- STATUS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- STATUS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- STATUS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskData
-
- STATUS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.TaskDataWrapper
-
- statusTracker() - Method in class org.apache.spark.api.java.JavaSparkContext
-
- statusTracker() - Method in class org.apache.spark.SparkContext
-
- StatusUpdate(String, long, Enumeration.Value, org.apache.spark.util.SerializableBuffer, int, Map<String, ResourceInformation>) - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- StatusUpdate - Class in org.apache.spark.scheduler.local
-
- StatusUpdate(long, Enumeration.Value, ByteBuffer) - Constructor for class org.apache.spark.scheduler.local.StatusUpdate
-
- StatusUpdate$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
-
- STD() - Static method in class org.apache.spark.ml.attribute.AttributeKeys
-
- std() - Method in class org.apache.spark.ml.attribute.NumericAttribute
-
- std() - Method in class org.apache.spark.ml.feature.StandardScalerModel
-
- std(Column, Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- std(Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- std() - Method in class org.apache.spark.mllib.feature.StandardScalerModel
-
- std() - Method in class org.apache.spark.mllib.random.LogNormalGenerator
-
- std(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: alias for stddev_samp
.
- stddev(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: alias for stddev_samp
.
- stddev(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: alias for stddev_samp
.
- stddev_pop(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the population standard deviation of
the expression in a group.
- stddev_pop(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the population standard deviation of
the expression in a group.
- stddev_samp(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sample standard deviation of
the expression in a group.
- stddev_samp(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sample standard deviation of
the expression in a group.
- stdev() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Compute the population standard deviation of this RDD's elements.
- stdev() - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the population standard deviation of this RDD's elements.
- stdev() - Method in class org.apache.spark.util.StatCounter
-
Return the population standard deviation of the values.
- stepSize() - Method in class org.apache.spark.ml.classification.FMClassificationModel
-
- stepSize() - Method in class org.apache.spark.ml.classification.FMClassifier
-
- stepSize() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- stepSize() - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- stepSize() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- stepSize() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- stepSize() - Method in class org.apache.spark.ml.feature.Word2Vec
-
- stepSize() - Method in class org.apache.spark.ml.feature.Word2VecModel
-
- stepSize() - Method in interface org.apache.spark.ml.param.shared.HasStepSize
-
Param for Step size to be used for each iteration of optimization (> 0).
- stepSize() - Method in class org.apache.spark.ml.regression.FMRegressionModel
-
- stepSize() - Method in class org.apache.spark.ml.regression.FMRegressor
-
- stepSize() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- stepSize() - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- stepSize() - Method in interface org.apache.spark.ml.tree.GBTParams
-
Param for Step size (a.k.a.
- stop() - Method in class org.apache.spark.api.java.JavaSparkContext
-
Shut down the SparkContext.
- stop() - Method in interface org.apache.spark.broadcast.BroadcastFactory
-
- stop() - Method in interface org.apache.spark.launcher.SparkAppHandle
-
Asks the application to stop.
- stop() - Method in interface org.apache.spark.metrics.sink.Sink
-
- stop() - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
- stop(int) - Method in interface org.apache.spark.scheduler.SchedulerBackend
-
- stop(int) - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- stop() - Method in class org.apache.spark.SparkContext
-
Shut down the SparkContext.
- stop(int) - Method in class org.apache.spark.SparkContext
-
Shut down the SparkContext with exit code that will passed to scheduler backend.
- stop() - Method in interface org.apache.spark.sql.connector.read.streaming.SparkDataStream
-
Stop this source and free any resources it has allocated.
- stop() - Method in class org.apache.spark.sql.SparkSession
-
Stop the underlying SparkContext
.
- stop() - Method in interface org.apache.spark.sql.streaming.StreamingQuery
-
Stops the execution of this query if it is running.
- stop() - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Stop the execution of the streams.
- stop(boolean) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Stop the execution of the streams.
- stop(boolean, boolean) - Method in class org.apache.spark.streaming.api.java.JavaStreamingContext
-
Deprecated.
Stop the execution of the streams.
- stop() - Method in class org.apache.spark.streaming.dstream.ConstantInputDStream
-
- stop() - Method in class org.apache.spark.streaming.dstream.InputDStream
-
Method called to stop receiving data.
- stop() - Method in class org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- stop(String) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Stop the receiver completely.
- stop(String, Throwable) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Stop the receiver completely due to an exception
- stop(boolean) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Stop the execution of the streams immediately (does not wait for all received data
to be processed).
- stop(boolean, boolean) - Method in class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Stop the execution of the streams, with option of ensuring all received data
has been processed.
- StopAllReceivers - Class in org.apache.spark.streaming.scheduler
-
This message will trigger ReceiverTrackerEndpoint to send stop signals to all registered
receivers.
- StopAllReceivers() - Constructor for class org.apache.spark.streaming.scheduler.StopAllReceivers
-
- StopBlockManagerMaster$() - Constructor for class org.apache.spark.storage.BlockManagerMessages.StopBlockManagerMaster$
-
- StopCoordinator - Class in org.apache.spark.scheduler
-
- StopCoordinator() - Constructor for class org.apache.spark.scheduler.StopCoordinator
-
- StopDriver$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopDriver$
-
- StopExecutor - Class in org.apache.spark.scheduler.local
-
- StopExecutor() - Constructor for class org.apache.spark.scheduler.local.StopExecutor
-
- StopExecutor$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutor$
-
- StopExecutors$() - Constructor for class org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutors$
-
- stopIndex() - Method in interface org.apache.spark.QueryContext
-
- StopMapOutputTracker - Class in org.apache.spark
-
- StopMapOutputTracker() - Constructor for class org.apache.spark.StopMapOutputTracker
-
- stopPosition() - Method in interface org.apache.spark.sql.avro.AvroUtils.RowReader
-
- StopReceiver - Class in org.apache.spark.streaming.receiver
-
- StopReceiver() - Constructor for class org.apache.spark.streaming.receiver.StopReceiver
-
- stopStandaloneSchedulerDriverEndpointError(Exception) - Static method in class org.apache.spark.errors.SparkCoreErrors
-
- stopWords() - Method in class org.apache.spark.ml.feature.StopWordsRemover
-
The words to be filtered out.
- StopWordsRemover - Class in org.apache.spark.ml.feature
-
A feature transformer that filters out stop words from input.
- StopWordsRemover(String) - Constructor for class org.apache.spark.ml.feature.StopWordsRemover
-
- StopWordsRemover() - Constructor for class org.apache.spark.ml.feature.StopWordsRemover
-
- STORAGE_LEVEL() - Static method in class org.apache.spark.ui.storage.ToolTips
-
- STORAGE_LEVEL_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDPartitionInfo
-
- STORAGE_LEVEL_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.RDDStorageInfo
-
- STORAGE_LEVEL_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StreamBlockData
-
- storageLevel() - Method in class org.apache.spark.sql.Dataset
-
Get the Dataset's current storage level, or StorageLevel.NONE if not persisted.
- storageLevel() - Method in class org.apache.spark.status.api.v1.RDDPartitionInfo
-
- storageLevel() - Method in class org.apache.spark.status.api.v1.RDDStorageInfo
-
- storageLevel() - Method in class org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
-
- storageLevel() - Method in class org.apache.spark.storage.BlockStatus
-
- storageLevel() - Method in class org.apache.spark.storage.BlockUpdatedInfo
-
- storageLevel() - Method in class org.apache.spark.storage.RDDInfo
-
- StorageLevel - Class in org.apache.spark.storage
-
:: DeveloperApi ::
Flags for controlling the storage of an RDD.
- StorageLevel() - Constructor for class org.apache.spark.storage.StorageLevel
-
- storageLevel() - Method in class org.apache.spark.streaming.receiver.Receiver
-
- storageLevelFromJson(JsonNode) - Static method in class org.apache.spark.util.JsonProtocol
-
- StorageLevels - Class in org.apache.spark.api.java
-
Expose some commonly useful storage level constants.
- StorageLevels() - Constructor for class org.apache.spark.api.java.StorageLevels
-
- storageLevelToJson(StorageLevel, JsonGenerator) - Static method in class org.apache.spark.util.JsonProtocol
-
- StorageUtils - Class in org.apache.spark.storage
-
Helper methods for storage-related objects.
- StorageUtils() - Constructor for class org.apache.spark.storage.StorageUtils
-
- store(T) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store a single item of received data to Spark's memory.
- store(ArrayBuffer<T>) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an ArrayBuffer of received data as a data block into Spark's memory.
- store(ArrayBuffer<T>, Object) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an ArrayBuffer of received data as a data block into Spark's memory.
- store(Iterator<T>) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>, Object) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>, Object) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(ByteBuffer) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store the bytes of received data as a data block into Spark's memory.
- store(ByteBuffer, Object) - Method in class org.apache.spark.streaming.receiver.Receiver
-
Store the bytes of received data as a data block into Spark's memory.
- storeBlock(StreamBlockId, ReceivedBlock) - Method in interface org.apache.spark.streaming.receiver.ReceivedBlockHandler
-
Store a received block with the given block id and return related metadata
- storedAsAndStoredByBothSpecifiedError(SqlBaseParser.CreateFileFormatContext) - Static method in class org.apache.spark.sql.errors.QueryParsingErrors
-
- StoreTypes - Class in org.apache.spark.status.protobuf
-
- StoreTypes.AccumulableInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.AccumulableInfo
- StoreTypes.AccumulableInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.AccumulableInfo
- StoreTypes.AccumulableInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ApplicationAttemptInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationAttemptInfo
- StoreTypes.ApplicationAttemptInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationAttemptInfo
- StoreTypes.ApplicationAttemptInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ApplicationEnvironmentInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationEnvironmentInfo
- StoreTypes.ApplicationEnvironmentInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationEnvironmentInfo
- StoreTypes.ApplicationEnvironmentInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ApplicationEnvironmentInfoWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationEnvironmentInfoWrapper
- StoreTypes.ApplicationEnvironmentInfoWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationEnvironmentInfoWrapper
- StoreTypes.ApplicationEnvironmentInfoWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ApplicationInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationInfo
- StoreTypes.ApplicationInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationInfo
- StoreTypes.ApplicationInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ApplicationInfoWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationInfoWrapper
- StoreTypes.ApplicationInfoWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ApplicationInfoWrapper
- StoreTypes.ApplicationInfoWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.AppSummary - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.AppSummary
- StoreTypes.AppSummary.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.AppSummary
- StoreTypes.AppSummaryOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.CachedQuantile - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.CachedQuantile
- StoreTypes.CachedQuantile.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.CachedQuantile
- StoreTypes.CachedQuantileOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.DeterministicLevel - Enum in org.apache.spark.status.protobuf
-
Protobuf enum org.apache.spark.status.protobuf.DeterministicLevel
- StoreTypes.ExecutorMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorMetrics
- StoreTypes.ExecutorMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorMetrics
- StoreTypes.ExecutorMetricsDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorMetricsDistributions
- StoreTypes.ExecutorMetricsDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorMetricsDistributions
- StoreTypes.ExecutorMetricsDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorPeakMetricsDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions
- StoreTypes.ExecutorPeakMetricsDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorPeakMetricsDistributions
- StoreTypes.ExecutorPeakMetricsDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorResourceRequest - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorResourceRequest
- StoreTypes.ExecutorResourceRequest.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorResourceRequest
- StoreTypes.ExecutorResourceRequestOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorStageSummary - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorStageSummary
- StoreTypes.ExecutorStageSummary.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorStageSummary
- StoreTypes.ExecutorStageSummaryOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorStageSummaryWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorStageSummaryWrapper
- StoreTypes.ExecutorStageSummaryWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorStageSummaryWrapper
- StoreTypes.ExecutorStageSummaryWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorSummary - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorSummary
- StoreTypes.ExecutorSummary.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorSummary
- StoreTypes.ExecutorSummaryOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ExecutorSummaryWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorSummaryWrapper
- StoreTypes.ExecutorSummaryWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ExecutorSummaryWrapper
- StoreTypes.ExecutorSummaryWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.InputMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.InputMetricDistributions
- StoreTypes.InputMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.InputMetricDistributions
- StoreTypes.InputMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.InputMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.InputMetrics
- StoreTypes.InputMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.InputMetrics
- StoreTypes.InputMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.JobData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.JobData
- StoreTypes.JobData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.JobData
- StoreTypes.JobDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.JobDataWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.JobDataWrapper
- StoreTypes.JobDataWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.JobDataWrapper
- StoreTypes.JobDataWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.JobExecutionStatus - Enum in org.apache.spark.status.protobuf
-
Protobuf enum org.apache.spark.status.protobuf.JobExecutionStatus
- StoreTypes.MemoryMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.MemoryMetrics
- StoreTypes.MemoryMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.MemoryMetrics
- StoreTypes.MemoryMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.OutputMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.OutputMetricDistributions
- StoreTypes.OutputMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.OutputMetricDistributions
- StoreTypes.OutputMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.OutputMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.OutputMetrics
- StoreTypes.OutputMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.OutputMetrics
- StoreTypes.OutputMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.PairStrings - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.PairStrings
- StoreTypes.PairStrings.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.PairStrings
- StoreTypes.PairStringsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.PoolData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.PoolData
- StoreTypes.PoolData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.PoolData
- StoreTypes.PoolDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ProcessSummary - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ProcessSummary
- StoreTypes.ProcessSummary.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ProcessSummary
- StoreTypes.ProcessSummaryOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ProcessSummaryWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ProcessSummaryWrapper
- StoreTypes.ProcessSummaryWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ProcessSummaryWrapper
- StoreTypes.ProcessSummaryWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDDataDistribution - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDDataDistribution
- StoreTypes.RDDDataDistribution.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDDataDistribution
- StoreTypes.RDDDataDistributionOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDOperationClusterWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationClusterWrapper
- StoreTypes.RDDOperationClusterWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationClusterWrapper
- StoreTypes.RDDOperationClusterWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDOperationEdge - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationEdge
- StoreTypes.RDDOperationEdge.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationEdge
- StoreTypes.RDDOperationEdgeOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDOperationGraphWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationGraphWrapper
- StoreTypes.RDDOperationGraphWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationGraphWrapper
- StoreTypes.RDDOperationGraphWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDOperationNode - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationNode
- StoreTypes.RDDOperationNode.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDOperationNode
- StoreTypes.RDDOperationNodeOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDPartitionInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDPartitionInfo
- StoreTypes.RDDPartitionInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDPartitionInfo
- StoreTypes.RDDPartitionInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDStorageInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDStorageInfo
- StoreTypes.RDDStorageInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDStorageInfo
- StoreTypes.RDDStorageInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RDDStorageInfoWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDStorageInfoWrapper
- StoreTypes.RDDStorageInfoWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RDDStorageInfoWrapper
- StoreTypes.RDDStorageInfoWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ResourceInformation - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceInformation
- StoreTypes.ResourceInformation.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceInformation
- StoreTypes.ResourceInformationOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ResourceProfileInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceProfileInfo
- StoreTypes.ResourceProfileInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceProfileInfo
- StoreTypes.ResourceProfileInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ResourceProfileWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceProfileWrapper
- StoreTypes.ResourceProfileWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ResourceProfileWrapper
- StoreTypes.ResourceProfileWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.RuntimeInfo - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RuntimeInfo
- StoreTypes.RuntimeInfo.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.RuntimeInfo
- StoreTypes.RuntimeInfoOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShufflePushReadMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions
- StoreTypes.ShufflePushReadMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShufflePushReadMetricDistributions
- StoreTypes.ShufflePushReadMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShufflePushReadMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShufflePushReadMetrics
- StoreTypes.ShufflePushReadMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShufflePushReadMetrics
- StoreTypes.ShufflePushReadMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShuffleReadMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleReadMetricDistributions
- StoreTypes.ShuffleReadMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleReadMetricDistributions
- StoreTypes.ShuffleReadMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShuffleReadMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleReadMetrics
- StoreTypes.ShuffleReadMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleReadMetrics
- StoreTypes.ShuffleReadMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShuffleWriteMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions
- StoreTypes.ShuffleWriteMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleWriteMetricDistributions
- StoreTypes.ShuffleWriteMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.ShuffleWriteMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleWriteMetrics
- StoreTypes.ShuffleWriteMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.ShuffleWriteMetrics
- StoreTypes.ShuffleWriteMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SinkProgress - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SinkProgress
- StoreTypes.SinkProgress.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SinkProgress
- StoreTypes.SinkProgressOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SourceProgress - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SourceProgress
- StoreTypes.SourceProgress.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SourceProgress
- StoreTypes.SourceProgressOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphClusterWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper
- StoreTypes.SparkPlanGraphClusterWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphClusterWrapper
- StoreTypes.SparkPlanGraphClusterWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphEdge - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphEdge
- StoreTypes.SparkPlanGraphEdge.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphEdge
- StoreTypes.SparkPlanGraphEdgeOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphNode - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphNode
- StoreTypes.SparkPlanGraphNode.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphNode
- StoreTypes.SparkPlanGraphNodeOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphNodeWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper
- StoreTypes.SparkPlanGraphNodeWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphNodeWrapper
- StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase - Enum in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphNodeWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SparkPlanGraphWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphWrapper
- StoreTypes.SparkPlanGraphWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SparkPlanGraphWrapper
- StoreTypes.SparkPlanGraphWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SpeculationStageSummary - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SpeculationStageSummary
- StoreTypes.SpeculationStageSummary.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SpeculationStageSummary
- StoreTypes.SpeculationStageSummaryOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SpeculationStageSummaryWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SpeculationStageSummaryWrapper
- StoreTypes.SpeculationStageSummaryWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SpeculationStageSummaryWrapper
- StoreTypes.SpeculationStageSummaryWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SQLExecutionUIData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SQLExecutionUIData
- StoreTypes.SQLExecutionUIData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SQLExecutionUIData
- StoreTypes.SQLExecutionUIDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.SQLPlanMetric - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SQLPlanMetric
- StoreTypes.SQLPlanMetric.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.SQLPlanMetric
- StoreTypes.SQLPlanMetricOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StageData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StageData
- StoreTypes.StageData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StageData
- StoreTypes.StageDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StageDataWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StageDataWrapper
- StoreTypes.StageDataWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StageDataWrapper
- StoreTypes.StageDataWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StageStatus - Enum in org.apache.spark.status.protobuf
-
Protobuf enum org.apache.spark.status.protobuf.StageStatus
- StoreTypes.StateOperatorProgress - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StateOperatorProgress
- StoreTypes.StateOperatorProgress.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StateOperatorProgress
- StoreTypes.StateOperatorProgressOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StreamBlockData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamBlockData
- StoreTypes.StreamBlockData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamBlockData
- StoreTypes.StreamBlockDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StreamingQueryData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryData
- StoreTypes.StreamingQueryData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryData
- StoreTypes.StreamingQueryDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StreamingQueryProgress - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryProgress
- StoreTypes.StreamingQueryProgress.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryProgress
- StoreTypes.StreamingQueryProgressOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.StreamingQueryProgressWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryProgressWrapper
- StoreTypes.StreamingQueryProgressWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.StreamingQueryProgressWrapper
- StoreTypes.StreamingQueryProgressWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.TaskData - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskData
- StoreTypes.TaskData.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskData
- StoreTypes.TaskDataOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.TaskDataWrapper - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskDataWrapper
- StoreTypes.TaskDataWrapper.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskDataWrapper
- StoreTypes.TaskDataWrapperOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.TaskMetricDistributions - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskMetricDistributions
- StoreTypes.TaskMetricDistributions.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskMetricDistributions
- StoreTypes.TaskMetricDistributionsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.TaskMetrics - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskMetrics
- StoreTypes.TaskMetrics.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskMetrics
- StoreTypes.TaskMetricsOrBuilder - Interface in org.apache.spark.status.protobuf
-
- StoreTypes.TaskResourceRequest - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskResourceRequest
- StoreTypes.TaskResourceRequest.Builder - Class in org.apache.spark.status.protobuf
-
Protobuf type org.apache.spark.status.protobuf.TaskResourceRequest
- StoreTypes.TaskResourceRequestOrBuilder - Interface in org.apache.spark.status.protobuf
-
- storeValue(T) - Method in class org.apache.spark.storage.memory.DeserializedValuesHolder
-
- storeValue(T) - Method in class org.apache.spark.storage.memory.SerializedValuesHolder
-
- storeValue(T) - Method in interface org.apache.spark.storage.memory.ValuesHolder
-
- str_to_map(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Creates a map after splitting the text into key/value pairs using delimiters.
- str_to_map(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Creates a map after splitting the text into key/value pairs using delimiters.
- str_to_map(Column) - Static method in class org.apache.spark.sql.functions
-
Creates a map after splitting the text into key/value pairs using delimiters.
- strategy() - Method in class org.apache.spark.ml.feature.Imputer
-
- strategy() - Method in class org.apache.spark.ml.feature.ImputerModel
-
- strategy() - Method in interface org.apache.spark.ml.feature.ImputerParams
-
The imputation strategy.
- Strategy - Class in org.apache.spark.mllib.tree.configuration
-
Stores all the configuration options for tree construction
param: algo Learning goal.
- Strategy(Enumeration.Value, Impurity, int, int, int, Enumeration.Value, Map<Object, Object>, int, double, int, double, boolean, int, double, boolean) - Constructor for class org.apache.spark.mllib.tree.configuration.Strategy
-
- Strategy(Enumeration.Value, Impurity, int, int, int, Enumeration.Value, Map<Object, Object>, int, double, int, double, boolean, int) - Constructor for class org.apache.spark.mllib.tree.configuration.Strategy
-
Backwards compatible constructor for
Strategy
- Strategy(Enumeration.Value, Impurity, int, int, int, Map<Integer, Integer>) - Constructor for class org.apache.spark.mllib.tree.configuration.Strategy
-
- StratifiedSamplingUtils - Class in org.apache.spark.util.random
-
Auxiliary functions and data structures for the sampleByKey method in PairRDDFunctions.
- StratifiedSamplingUtils() - Constructor for class org.apache.spark.util.random.StratifiedSamplingUtils
-
- STREAM() - Static method in class org.apache.spark.storage.BlockId
-
- StreamBlockId - Class in org.apache.spark.storage
-
- StreamBlockId(int, long) - Constructor for class org.apache.spark.storage.StreamBlockId
-
- streamedOperatorUnsupportedByDataSourceError(String, String) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- streamId() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- streamId() - Method in class org.apache.spark.storage.StreamBlockId
-
- streamId() - Method in class org.apache.spark.streaming.receiver.Receiver
-
Get the unique identifier the receiver input stream that this
receiver is associated with.
- streamId() - Method in class org.apache.spark.streaming.scheduler.ReceiverInfo
-
- streamIdToInputInfo() - Method in class org.apache.spark.streaming.scheduler.BatchInfo
-
- StreamingConf - Class in org.apache.spark.streaming
-
- StreamingConf() - Constructor for class org.apache.spark.streaming.StreamingConf
-
- StreamingContext - Class in org.apache.spark.streaming
-
- StreamingContext(SparkContext, Duration) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Create a StreamingContext using an existing SparkContext.
- StreamingContext(SparkConf, Duration) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Create a StreamingContext by providing the configuration necessary for a new SparkContext.
- StreamingContext(String, String, Duration, String, Seq<String>, Map<String, String>) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Create a StreamingContext by providing the details necessary for creating a new SparkContext.
- StreamingContext(String, Configuration) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Recreate a StreamingContext from a checkpoint file.
- StreamingContext(String) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Recreate a StreamingContext from a checkpoint file.
- StreamingContext(String, SparkContext) - Constructor for class org.apache.spark.streaming.StreamingContext
-
Deprecated.
Recreate a StreamingContext from a checkpoint file using an existing SparkContext.
- StreamingContextPythonHelper - Class in org.apache.spark.streaming
-
- StreamingContextPythonHelper() - Constructor for class org.apache.spark.streaming.StreamingContextPythonHelper
-
- StreamingContextState - Enum in org.apache.spark.streaming
-
:: DeveloperApi ::
Represents the state of a StreamingContext.
- StreamingDataWriterFactory - Interface in org.apache.spark.sql.connector.write.streaming
-
- streamingIntoViewNotSupportedError(String) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- StreamingKMeans - Class in org.apache.spark.mllib.clustering
-
StreamingKMeans provides methods for configuring a
streaming k-means analysis, training the model on streaming,
and using the model to make predictions on streaming data.
- StreamingKMeans(int, double, String) - Constructor for class org.apache.spark.mllib.clustering.StreamingKMeans
-
- StreamingKMeans() - Constructor for class org.apache.spark.mllib.clustering.StreamingKMeans
-
- StreamingKMeansModel - Class in org.apache.spark.mllib.clustering
-
StreamingKMeansModel extends MLlib's KMeansModel for streaming
algorithms, so it can keep track of a continuously updated weight
associated with each cluster, and also update the model by
doing a single iteration of the standard k-means algorithm.
- StreamingKMeansModel(Vector[], double[]) - Constructor for class org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- StreamingLinearAlgorithm<M extends GeneralizedLinearModel,A extends GeneralizedLinearAlgorithm<M>> - Class in org.apache.spark.mllib.regression
-
StreamingLinearAlgorithm implements methods for continuously
training a generalized linear model on streaming data,
and using it for prediction on (possibly different) streaming data.
- StreamingLinearAlgorithm() - Constructor for class org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
- StreamingLinearRegressionWithSGD - Class in org.apache.spark.mllib.regression
-
Train or predict a linear regression model on streaming data.
- StreamingLinearRegressionWithSGD() - Constructor for class org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Construct a StreamingLinearRegression object with default parameters:
{stepSize: 0.1, numIterations: 50, miniBatchFraction: 1.0}.
- StreamingListener - Interface in org.apache.spark.streaming.scheduler
-
:: DeveloperApi ::
A listener interface for receiving information about an ongoing streaming
computation.
- StreamingListenerBatchCompleted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerBatchCompleted(BatchInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- StreamingListenerBatchStarted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerBatchStarted(BatchInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- StreamingListenerBatchSubmitted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerBatchSubmitted(BatchInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- StreamingListenerEvent - Interface in org.apache.spark.streaming.scheduler
-
:: DeveloperApi ::
Base trait for events related to StreamingListener
- StreamingListenerOutputOperationCompleted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerOutputOperationCompleted(OutputOperationInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- StreamingListenerOutputOperationStarted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerOutputOperationStarted(OutputOperationInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- StreamingListenerReceiverError - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerReceiverError(ReceiverInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- StreamingListenerReceiverStarted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerReceiverStarted(ReceiverInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- StreamingListenerReceiverStopped - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerReceiverStopped(ReceiverInfo) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- StreamingListenerStreamingStarted - Class in org.apache.spark.streaming.scheduler
-
- StreamingListenerStreamingStarted(long) - Constructor for class org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- StreamingLogisticRegressionWithSGD - Class in org.apache.spark.mllib.classification
-
Train or predict a logistic regression model on streaming data.
- StreamingLogisticRegressionWithSGD() - Constructor for class org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Construct a StreamingLogisticRegression object with default parameters:
{stepSize: 0.1, numIterations: 50, miniBatchFraction: 1.0, regParam: 0.0}.
- StreamingQuery - Interface in org.apache.spark.sql.streaming
-
A handle to a query that is executing continuously in the background as new data arrives.
- StreamingQueryException - Exception in org.apache.spark.sql.streaming
-
- StreamingQueryException(String, Throwable, String, String, String, Map<String, String>) - Constructor for exception org.apache.spark.sql.streaming.StreamingQueryException
-
- StreamingQueryListener - Class in org.apache.spark.sql.streaming
-
- StreamingQueryListener() - Constructor for class org.apache.spark.sql.streaming.StreamingQueryListener
-
- StreamingQueryListener.Event - Interface in org.apache.spark.sql.streaming
-
- StreamingQueryListener.QueryIdleEvent - Class in org.apache.spark.sql.streaming
-
Event representing that query is idle and waiting for new data to process.
- StreamingQueryListener.QueryProgressEvent - Class in org.apache.spark.sql.streaming
-
Event representing any progress updates in a query.
- StreamingQueryListener.QueryStartedEvent - Class in org.apache.spark.sql.streaming
-
Event representing the start of a query
param: id A unique query id that persists across restarts.
- StreamingQueryListener.QueryTerminatedEvent - Class in org.apache.spark.sql.streaming
-
Event representing that termination of a query.
- StreamingQueryManager - Class in org.apache.spark.sql.streaming
-
- StreamingQueryProgress - Class in org.apache.spark.sql.streaming
-
Information about progress made in the execution of a
StreamingQuery
during
a trigger.
- StreamingQueryProgressSerializer - Class in org.apache.spark.status.protobuf.sql
-
- StreamingQueryProgressSerializer() - Constructor for class org.apache.spark.status.protobuf.sql.StreamingQueryProgressSerializer
-
- StreamingQueryStatus - Class in org.apache.spark.sql.streaming
-
Reports information about the instantaneous status of a streaming query.
- streamingSourcesDoNotSupportCommonExecutionModeError(Seq<String>, Seq<String>) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- StreamingStatistics - Class in org.apache.spark.status.api.v1.streaming
-
- StreamingTest - Class in org.apache.spark.mllib.stat.test
-
Performs online 2-sample significance testing for a stream of (Boolean, Double) pairs.
- StreamingTest() - Constructor for class org.apache.spark.mllib.stat.test.StreamingTest
-
- StreamingTestMethod - Interface in org.apache.spark.mllib.stat.test
-
- StreamingWrite - Interface in org.apache.spark.sql.connector.write.streaming
-
An interface that defines how to write the data to data source in streaming queries.
- StreamInputInfo - Class in org.apache.spark.streaming.scheduler
-
:: DeveloperApi ::
Track the information of input stream at specified batch time.
- StreamInputInfo(int, long, Map<String, Object>) - Constructor for class org.apache.spark.streaming.scheduler.StreamInputInfo
-
- streamJoinStreamWithoutEqualityPredicateUnsupportedError(LogicalPlan) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- streamName() - Method in class org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- streams() - Method in class org.apache.spark.sql.SparkSession
-
Returns a StreamingQueryManager
that allows managing all the
StreamingQuery
s active on this
.
- streams() - Method in class org.apache.spark.sql.SQLContext
-
- StreamSinkProvider - Interface in org.apache.spark.sql.sources
-
::Experimental::
Implemented by objects that can produce a streaming Sink
for a specific format or system.
- StreamSourceProvider - Interface in org.apache.spark.sql.sources
-
::Experimental::
Implemented by objects that can produce a streaming Source
for a specific format or system.
- STRING() - Static method in class org.apache.spark.api.r.SerializationFormats
-
- string() - Method in class org.apache.spark.sql.ColumnName
-
Creates a new StructField
of type string.
- STRING() - Static method in class org.apache.spark.sql.Encoders
-
An encoder for nullable string type.
- StringArrayParam - Class in org.apache.spark.ml.param
-
Specialized version of Param[Array[String}
for Java.
- StringArrayParam(Params, String, String, Function1<String[], Object>) - Constructor for class org.apache.spark.ml.param.StringArrayParam
-
- StringArrayParam(Params, String, String) - Constructor for class org.apache.spark.ml.param.StringArrayParam
-
- StringContains - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff the attribute evaluates to
a string that contains the string value
.
- StringContains(String, String) - Constructor for class org.apache.spark.sql.sources.StringContains
-
- StringEndsWith - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff the attribute evaluates to
a string that ends with value
.
- StringEndsWith(String, String) - Constructor for class org.apache.spark.sql.sources.StringEndsWith
-
- stringHalfWidth(String) - Static method in class org.apache.spark.util.Utils
-
Return the number of half widths in a given string.
- StringIndexer - Class in org.apache.spark.ml.feature
-
A label indexer that maps string column(s) of labels to ML column(s) of label indices.
- StringIndexer(String) - Constructor for class org.apache.spark.ml.feature.StringIndexer
-
- StringIndexer() - Constructor for class org.apache.spark.ml.feature.StringIndexer
-
- StringIndexerAggregator - Class in org.apache.spark.ml.feature
-
A SQL Aggregator
used by StringIndexer
to count labels in string columns during fitting.
- StringIndexerAggregator(int) - Constructor for class org.apache.spark.ml.feature.StringIndexerAggregator
-
- StringIndexerBase - Interface in org.apache.spark.ml.feature
-
- StringIndexerModel - Class in org.apache.spark.ml.feature
-
- StringIndexerModel(String, String[][]) - Constructor for class org.apache.spark.ml.feature.StringIndexerModel
-
- StringIndexerModel(String, String[]) - Constructor for class org.apache.spark.ml.feature.StringIndexerModel
-
- StringIndexerModel(String[]) - Constructor for class org.apache.spark.ml.feature.StringIndexerModel
-
- StringIndexerModel(String[][]) - Constructor for class org.apache.spark.ml.feature.StringIndexerModel
-
- stringIndexerOrderType() - Method in class org.apache.spark.ml.feature.RFormula
-
- stringIndexerOrderType() - Method in interface org.apache.spark.ml.feature.RFormulaBase
-
Param for how to order categories of a string FEATURE column used by StringIndexer
.
- stringIndexerOrderType() - Method in class org.apache.spark.ml.feature.RFormulaModel
-
- stringOrderType() - Method in class org.apache.spark.ml.feature.StringIndexer
-
- stringOrderType() - Method in interface org.apache.spark.ml.feature.StringIndexerBase
-
Param for how to order labels of string column.
- stringOrderType() - Method in class org.apache.spark.ml.feature.StringIndexerModel
-
- StringRRDD<T> - Class in org.apache.spark.api.r
-
An RDD that stores R objects as Array[String].
- StringRRDD(RDD<T>, byte[], String, byte[], Object[], ClassTag<T>) - Constructor for class org.apache.spark.api.r.StringRRDD
-
- StringStartsWith - Class in org.apache.spark.sql.sources
-
A filter that evaluates to true
iff the attribute evaluates to
a string that starts with value
.
- StringStartsWith(String, String) - Constructor for class org.apache.spark.sql.sources.StringStartsWith
-
- StringToColumn(StringContext) - Constructor for class org.apache.spark.sql.SQLImplicits.StringToColumn
-
- stringToField() - Static method in class org.apache.spark.sql.types.DayTimeIntervalType
-
- stringToField() - Static method in class org.apache.spark.sql.types.YearMonthIntervalType
-
- stringToSeq(String) - Static method in class org.apache.spark.util.Utils
-
- StringType - Static variable in class org.apache.spark.sql.types.DataTypes
-
Gets the StringType object.
- StringType - Class in org.apache.spark.sql.types
-
The data type representing String
values.
- StringType() - Constructor for class org.apache.spark.sql.types.StringType
-
- StringTypeExpression - Class in org.apache.spark.sql.types
-
- StringTypeExpression() - Constructor for class org.apache.spark.sql.types.StringTypeExpression
-
- stripDollars(String) - Static method in class org.apache.spark.util.Utils
-
Remove trailing dollar signs from qualified class name,
and return the trailing part after the last dollar sign in the middle
- stronglyConnectedComponents(int) - Method in class org.apache.spark.graphx.GraphOps
-
Compute the strongly connected component (SCC) of each vertex and return a graph with the
vertex value containing the lowest vertex id in the SCC containing that vertex.
- StronglyConnectedComponents - Class in org.apache.spark.graphx.lib
-
Strongly connected components algorithm implementation.
- StronglyConnectedComponents() - Constructor for class org.apache.spark.graphx.lib.StronglyConnectedComponents
-
- struct(Seq<StructField>) - Method in class org.apache.spark.sql.ColumnName
-
Creates a new StructField
of type struct.
- struct(StructType) - Method in class org.apache.spark.sql.ColumnName
-
Creates a new StructField
of type struct.
- struct(Column...) - Static method in class org.apache.spark.sql.functions
-
Creates a new struct column.
- struct(String, String...) - Static method in class org.apache.spark.sql.functions
-
Creates a new struct column that composes multiple input columns.
- struct(Seq<Column>) - Static method in class org.apache.spark.sql.functions
-
Creates a new struct column.
- struct(String, Seq<String>) - Static method in class org.apache.spark.sql.functions
-
Creates a new struct column that composes multiple input columns.
- StructField - Class in org.apache.spark.sql.types
-
A field inside a StructType.
- StructField(String, DataType, boolean, Metadata) - Constructor for class org.apache.spark.sql.types.StructField
-
- StructType - Class in org.apache.spark.sql.types
-
- StructType(StructField[]) - Constructor for class org.apache.spark.sql.types.StructType
-
- StructType() - Constructor for class org.apache.spark.sql.types.StructType
-
No-arg constructor for kryo.
- structTypeToV2Columns(StructType) - Static method in class org.apache.spark.sql.connector.catalog.CatalogV2Util
-
Converts a StructType to DS v2 columns, which decodes the StructField metadata to v2 column
comment and default value or generation expression.
- stsCredentials(String, String) - Method in class org.apache.spark.streaming.kinesis.SparkAWSCredentials.Builder
-
Use STS to assume an IAM role for temporary session-based authentication.
- stsCredentials(String, String, String) - Method in class org.apache.spark.streaming.kinesis.SparkAWSCredentials.Builder
-
Use STS to assume an IAM role for temporary session-based authentication.
- StudentTTest - Class in org.apache.spark.mllib.stat.test
-
Performs Students's 2-sample t-test.
- StudentTTest() - Constructor for class org.apache.spark.mllib.stat.test.StudentTTest
-
- subClass() - Method in class org.apache.spark.ErrorInfo
-
- subgraph(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - Method in class org.apache.spark.graphx.Graph
-
Restricts the graph to only the vertices and edges satisfying the predicates.
- subgraph(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - Method in class org.apache.spark.graphx.impl.GraphImpl
-
- SUBMISSION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.JobData
-
- SUBMISSION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData
-
- SUBMISSION_TIME_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.StageData
-
- submissionTime() - Method in class org.apache.spark.scheduler.StageInfo
-
When this stage was submitted from the DAGScheduler to a TaskScheduler.
- submissionTime() - Method in interface org.apache.spark.SparkStageInfo
-
- submissionTime() - Method in class org.apache.spark.SparkStageInfoImpl
-
- submissionTime() - Method in class org.apache.spark.status.api.v1.JobData
-
- submissionTime() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
-
- submissionTime() - Method in class org.apache.spark.status.api.v1.StageData
-
- submissionTime() - Method in class org.apache.spark.status.LiveJob
-
- submissionTime() - Method in class org.apache.spark.streaming.scheduler.BatchInfo
-
- submitJob(RDD<T>, Function1<Iterator<T>, U>, Seq<Object>, Function2<Object, U, BoxedUnit>, Function0<R>) - Method in interface org.apache.spark.JobSubmitter
-
Submit a job for execution and return a FutureAction holding the result.
- submitJob(RDD<T>, Function1<Iterator<T>, U>, Seq<Object>, Function2<Object, U, BoxedUnit>, Function0<R>) - Method in class org.apache.spark.SparkContext
-
Submit a job for execution and return a FutureJob holding the result.
- submitTasks(TaskSet) - Method in interface org.apache.spark.scheduler.TaskScheduler
-
- subModels() - Method in class org.apache.spark.ml.tuning.CrossValidatorModel
-
- subModels() - Method in class org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- subprocessExitedError(int, org.apache.spark.util.CircularBuffer, Throwable) - Static method in class org.apache.spark.sql.errors.QueryExecutionErrors
-
- subqueryNotAllowedInMergeCondition(String, Expression) - Static method in class org.apache.spark.sql.errors.QueryCompilationErrors
-
- subsamplingRate() - Method in class org.apache.spark.ml.classification.GBTClassificationModel
-
- subsamplingRate() - Method in class org.apache.spark.ml.classification.GBTClassifier
-
- subsamplingRate() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
- subsamplingRate() - Method in class org.apache.spark.ml.classification.RandomForestClassifier
-
- subsamplingRate() - Method in class org.apache.spark.ml.clustering.LDA
-
- subsamplingRate() - Method in class org.apache.spark.ml.clustering.LDAModel
-
- subsamplingRate() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
For Online optimizer only: optimizer
= "online".
- subsamplingRate() - Method in class org.apache.spark.ml.regression.GBTRegressionModel
-
- subsamplingRate() - Method in class org.apache.spark.ml.regression.GBTRegressor
-
- subsamplingRate() - Method in class org.apache.spark.ml.regression.RandomForestRegressionModel
-
- subsamplingRate() - Method in class org.apache.spark.ml.regression.RandomForestRegressor
-
- subsamplingRate() - Method in interface org.apache.spark.ml.tree.TreeEnsembleParams
-
Fraction of the training data used for learning each decision tree, in range (0, 1].
- subsamplingRate() - Method in class org.apache.spark.mllib.tree.configuration.Strategy
-
- subsetAccuracy() - Method in class org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns subset accuracy
(for equal sets of labels)
- substituteAppId(String, String) - Static method in class org.apache.spark.util.Utils
-
Replaces all the {{APP_ID}} occurrences with the App Id.
- substituteAppNExecIds(String, String, String) - Static method in class org.apache.spark.util.Utils
-
Replaces all the {{EXECUTOR_ID}} occurrences with the Executor Id
and {{APP_ID}} occurrences with the App Id.
- substr(Column, Column) - Method in class org.apache.spark.sql.Column
-
An expression that returns a substring.
- substr(int, int) - Method in class org.apache.spark.sql.Column
-
An expression that returns a substring.
- substr(Column, Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns the substring of str
that starts at pos
and is of length len
,
or the slice of byte array that starts at pos
and is of length len
.
- substr(Column, Column) - Static method in class org.apache.spark.sql.functions
-
Returns the substring of str
that starts at pos
,
or the slice of byte array that starts at pos
.
- substring(Column, int, int) - Static method in class org.apache.spark.sql.functions
-
Substring starts at pos
and is of length len
when str is String type or
returns the slice of byte array that starts at pos
in byte and is of length len
when str is Binary type
- substring_index(Column, String, int) - Static method in class org.apache.spark.sql.functions
-
Returns the substring from string str before count occurrences of the delimiter delim.
- subtract(JavaDoubleRDD) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaDoubleRDD, int) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaDoubleRDD, Partitioner) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaPairRDD<K, V>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaPairRDD<K, V>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaPairRDD<K, V>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaRDD<T>) - Method in class org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaRDD<T>, int) - Method in class org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(JavaRDD<T>, Partitioner) - Method in class org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(Term) - Static method in class org.apache.spark.ml.feature.Dot
-
- subtract(Term) - Static method in class org.apache.spark.ml.feature.EmptyTerm
-
- subtract(Term) - Method in interface org.apache.spark.ml.feature.Term
-
Fold by adding deletion terms to the left.
- subtract(BlockMatrix) - Method in class org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Subtracts the given block matrix other
from this
block matrix: this - other
.
- subtract(RDD<T>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(RDD<T>, int) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(RDD<T>, Partitioner, Ordering<T>) - Method in class org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this
that are not in other
.
- subtract(long, long) - Static method in class org.apache.spark.streaming.util.RawTextHelper
-
- subtractByKey(JavaPairRDD<K, W>) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractByKey(JavaPairRDD<K, W>, int) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractByKey(JavaPairRDD<K, W>, Partitioner) - Method in class org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractByKey(RDD<Tuple2<K, W>>, ClassTag<W>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractByKey(RDD<Tuple2<K, W>>, int, ClassTag<W>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractByKey(RDD<Tuple2<K, W>>, Partitioner, ClassTag<W>) - Method in class org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this
whose keys are not in other
.
- subtractMetrics(TaskMetrics, TaskMetrics) - Static method in class org.apache.spark.status.LiveEntityHelpers
-
Subtract m2 values from m1.
- SUCCEEDED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsDistributions
-
- SUCCEEDED_TASKS_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ExecutorStageSummary
-
- succeededTasks() - Method in class org.apache.spark.status.api.v1.ExecutorMetricsDistributions
-
- succeededTasks() - Method in class org.apache.spark.status.api.v1.ExecutorStageSummary
-
- succeededTasks() - Method in class org.apache.spark.status.LiveExecutorStageSummary
-
- Success() - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- success(T) - Static method in class org.apache.spark.ml.feature.RFormulaParser
-
- Success - Class in org.apache.spark
-
:: DeveloperApi ::
Task succeeded.
- Success() - Constructor for class org.apache.spark.Success
-
- successful() - Method in class org.apache.spark.scheduler.TaskInfo
-
- successJobIds() - Method in class org.apache.spark.status.api.v1.sql.ExecutionData
-
- sum() - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Add up the elements in this RDD.
- sum(Column, Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- sum(Column) - Static method in class org.apache.spark.ml.stat.Summarizer
-
- Sum() - Static method in class org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- sum() - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Add up the elements in this RDD.
- Sum - Class in org.apache.spark.sql.connector.expressions.aggregate
-
An aggregate function that returns the summation of all the values in a group.
- Sum(Expression, boolean) - Constructor for class org.apache.spark.sql.connector.expressions.aggregate.Sum
-
- sum(MapFunction<T, Double>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
-
Deprecated.
Sum aggregate function for floating point (double) type.
- sum(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
-
Deprecated.
Sum aggregate function for floating point (double) type.
- sum(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sum of all values in the expression.
- sum(String) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sum of all values in the given column.
- sum(String...) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute the sum for each numeric columns for each group.
- sum(Seq<String>) - Method in class org.apache.spark.sql.RelationalGroupedDataset
-
Compute the sum for each numeric columns for each group.
- sum() - Method in class org.apache.spark.util.DoubleAccumulator
-
Returns the sum of elements added to the accumulator.
- sum() - Method in class org.apache.spark.util.LongAccumulator
-
Returns the sum of elements added to the accumulator.
- sum() - Method in class org.apache.spark.util.StatCounter
-
- sum_distinct(Column) - Static method in class org.apache.spark.sql.functions
-
Aggregate function: returns the sum of distinct values in the expression.
- sumApprox(long, Double) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Approximate operation to return the sum within a timeout.
- sumApprox(long) - Method in class org.apache.spark.api.java.JavaDoubleRDD
-
Approximate operation to return the sum within a timeout.
- sumApprox(long, double) - Method in class org.apache.spark.rdd.DoubleRDDFunctions
-
Approximate operation to return the sum within a timeout.
- sumDistinct(Column) - Static method in class org.apache.spark.sql.functions
-
- sumDistinct(String) - Static method in class org.apache.spark.sql.functions
-
- sumLong(MapFunction<T, Long>) - Static method in class org.apache.spark.sql.expressions.javalang.typed
-
Deprecated.
Sum aggregate function for integral (long, i.e.
- sumLong(Function1<IN, Object>) - Static method in class org.apache.spark.sql.expressions.scalalang.typed
-
Deprecated.
Sum aggregate function for integral (long, i.e.
- Summarizer - Class in org.apache.spark.ml.stat
-
Tools for vectorized statistics on MLlib Vectors.
- Summarizer() - Constructor for class org.apache.spark.ml.stat.Summarizer
-
- summary() - Method in class org.apache.spark.ml.classification.FMClassificationModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.classification.LinearSVCModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.classification.LogisticRegressionModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.classification.RandomForestClassificationModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.clustering.BisectingKMeansModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.clustering.GaussianMixtureModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.clustering.KMeansModel
-
Gets summary of model on training set.
- summary() - Method in class org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
Gets R-like summary of model on training set.
- summary() - Method in class org.apache.spark.ml.regression.LinearRegressionModel
-
Gets summary (e.g.
- summary(Column, Column) - Method in class org.apache.spark.ml.stat.SummaryBuilder
-
Returns an aggregate object that contains the summary of the column with the requested metrics.
- summary(Column) - Method in class org.apache.spark.ml.stat.SummaryBuilder
-
- summary() - Method in interface org.apache.spark.ml.util.HasTrainingSummary
-
Gets summary of model on training set.
- summary(String...) - Method in class org.apache.spark.sql.Dataset
-
Computes specified statistics for numeric and string columns.
- summary(Seq<String>) - Method in class org.apache.spark.sql.Dataset
-
Computes specified statistics for numeric and string columns.
- SummaryBuilder - Class in org.apache.spark.ml.stat
-
A builder object that provides summary statistics about a given column.
- SummaryBuilder() - Constructor for class org.apache.spark.ml.stat.SummaryBuilder
-
- sumMat() - Method in class org.apache.spark.ml.clustering.KMeansAggregator
-
- supportColumnarReads(InputPartition) - Method in interface org.apache.spark.sql.connector.read.PartitionReaderFactory
-
Returns true if the given
InputPartition
should be read by Spark in a columnar way.
- supportCompletePushDown(Aggregation) - Method in interface org.apache.spark.sql.connector.read.SupportsPushDownAggregates
-
Whether the datasource support complete aggregation push-down.
- supportedCustomMetrics() - Method in interface org.apache.spark.sql.connector.read.Scan
-
Returns an array of supported custom metrics with name and description.
- supportedCustomMetrics() - Method in interface org.apache.spark.sql.connector.write.Write
-
Returns an array of supported custom metrics with name and description.
- supportedFeatureSubsetStrategies() - Static method in class org.apache.spark.ml.classification.RandomForestClassifier
-
Accessor for supported featureSubsetStrategy settings: auto, all, onethird, sqrt, log2
- supportedFeatureSubsetStrategies() - Static method in class org.apache.spark.ml.regression.RandomForestRegressor
-
Accessor for supported featureSubsetStrategy settings: auto, all, onethird, sqrt, log2
- supportedFeatureSubsetStrategies() - Static method in class org.apache.spark.mllib.tree.RandomForest
-
List of supported feature subset sampling strategies.
- supportedImpurities() - Static method in class org.apache.spark.ml.classification.DecisionTreeClassifier
-
Accessor for supported impurities: entropy, gini
- supportedImpurities() - Static method in class org.apache.spark.ml.classification.RandomForestClassifier
-
Accessor for supported impurity settings: entropy, gini
- supportedImpurities() - Static method in class org.apache.spark.ml.regression.DecisionTreeRegressor
-
Accessor for supported impurities: variance
- supportedImpurities() - Static method in class org.apache.spark.ml.regression.RandomForestRegressor
-
Accessor for supported impurity settings: variance
- supportedLossTypes() - Static method in class org.apache.spark.ml.classification.GBTClassifier
-
Accessor for supported loss settings: logistic
- supportedLossTypes() - Static method in class org.apache.spark.ml.regression.GBTRegressor
-
Accessor for supported loss settings: squared (L2), absolute (L1)
- supportedOptimizers() - Method in class org.apache.spark.ml.clustering.LDA
-
- supportedOptimizers() - Method in class org.apache.spark.ml.clustering.LDAModel
-
- supportedOptimizers() - Method in interface org.apache.spark.ml.clustering.LDAParams
-
Supported values for Param optimizer
.
- supportedSelectorTypes() - Static method in class org.apache.spark.mllib.feature.ChiSqSelector
-
Set of selector types that ChiSqSelector supports.
- SupportsAdmissionControl - Interface in org.apache.spark.sql.connector.read.streaming
-
A mix-in interface for
SparkDataStream
streaming sources to signal that they can control
the rate of data ingested into the system.
- SupportsAtomicPartitionManagement - Interface in org.apache.spark.sql.connector.catalog
-
An atomic partition interface of
Table
to operate multiple partitions atomically.
- SupportsCatalogOptions - Interface in org.apache.spark.sql.connector.catalog
-
An interface, which TableProviders can implement, to support table existence checks and creation
through a catalog, without having to use table identifiers.
- supportsColumnarInput(Seq<Attribute>) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Can convertColumnarBatchToCachedBatch()
be called instead of
convertInternalRowToCachedBatch()
for this given schema? True if it can and false if it
cannot.
- supportsColumnarOutput(StructType) - Method in interface org.apache.spark.sql.columnar.CachedBatchSerializer
-
Can convertCachedBatchToColumnarBatch()
be called instead of
convertCachedBatchToInternalRow()
for this given schema? True if it can and false if it
cannot.
- supportsDataType(DataType) - Static method in class org.apache.spark.sql.avro.AvroUtils
-
- SupportsDelete - Interface in org.apache.spark.sql.connector.catalog
-
A mix-in interface for
Table
delete support.
- SupportsDeleteV2 - Interface in org.apache.spark.sql.connector.catalog
-
A mix-in interface for
Table
delete support.
- SupportsDelta - Interface in org.apache.spark.sql.connector.write
-
- SupportsDynamicOverwrite - Interface in org.apache.spark.sql.connector.write
-
Write builder trait for tables that support dynamic partition overwrite.
- supportsExternalMetadata() - Method in interface org.apache.spark.sql.connector.catalog.TableProvider
-
Returns true if the source has the ability of accepting external table metadata when getting
tables.
- SupportsIndex - Interface in org.apache.spark.sql.connector.catalog.index
-
Table methods for working with index
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- supportsLimit() - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns ture if dialect supports LIMIT clause.
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- supportsLimit() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- SupportsMetadataColumns - Interface in org.apache.spark.sql.connector.catalog
-
An interface for exposing data columns for a table that are not in the table schema.
- SupportsNamespaces - Interface in org.apache.spark.sql.connector.catalog
-
Catalog methods for working with namespaces.
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- supportsOffset() - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
Returns ture if dialect supports OFFSET clause.
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- supportsOffset() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- SupportsOverwrite - Interface in org.apache.spark.sql.connector.write
-
Write builder trait for tables that support overwrite by filter.
- SupportsOverwriteV2 - Interface in org.apache.spark.sql.connector.write
-
Write builder trait for tables that support overwrite by filter.
- SupportsPartitionManagement - Interface in org.apache.spark.sql.connector.catalog
-
A partition interface of
Table
.
- SupportsPushDownAggregates - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownFilters - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownLimit - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownOffset - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownRequiredColumns - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownTableSample - Interface in org.apache.spark.sql.connector.read
-
A mix-in interface for
Scan
.
- SupportsPushDownTopN - Interface in org.apache.spark.sql.connector.read
-
- SupportsPushDownV2Filters - Interface in org.apache.spark.sql.connector.read
-
- SupportsRead - Interface in org.apache.spark.sql.connector.catalog
-
A mix-in interface of
Table
, to indicate that it's readable.
- supportsReliableStorage() - Method in interface org.apache.spark.shuffle.api.ShuffleDriverComponents
-
Does this shuffle component support reliable storage - external to the lifecycle of the
executor host ? For example, writing shuffle data to a distributed filesystem or
persisting it in a remote shuffle service.
- SupportsReportOrdering - Interface in org.apache.spark.sql.connector.read
-
A mix in interface for
Scan
.
- SupportsReportPartitioning - Interface in org.apache.spark.sql.connector.read
-
A mix in interface for
Scan
.
- SupportsReportStatistics - Interface in org.apache.spark.sql.connector.read
-
A mix in interface for
Scan
.
- SupportsRowLevelOperations - Interface in org.apache.spark.sql.connector.catalog
-
A mix-in interface for
Table
row-level operations support.
- SupportsRuntimeFiltering - Interface in org.apache.spark.sql.connector.read
-
A mix-in interface for
Scan
.
- SupportsRuntimeV2Filtering - Interface in org.apache.spark.sql.connector.read
-
A mix-in interface for
Scan
.
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.DB2Dialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.DerbyDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.H2Dialect
-
- supportsTableSample() - Method in class org.apache.spark.sql.jdbc.JdbcDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.MySQLDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.NoopDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.OracleDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.PostgresDialect
-
- supportsTableSample() - Static method in class org.apache.spark.sql.jdbc.TeradataDialect
-
- SupportsTriggerAvailableNow - Interface in org.apache.spark.sql.connector.read.streaming
-
An interface for streaming sources that supports running in Trigger.AvailableNow mode, which
will process all the available data at the beginning of the query in (possibly) multiple batches.
- SupportsTruncate - Interface in org.apache.spark.sql.connector.write
-
Write builder trait for tables that support truncation.
- SupportsWrite - Interface in org.apache.spark.sql.connector.catalog
-
A mix-in interface of
Table
, to indicate that it's writable.
- surrogateDF() - Method in class org.apache.spark.ml.feature.ImputerModel
-
- SVDPlusPlus - Class in org.apache.spark.graphx.lib
-
Implementation of SVD++ algorithm.
- SVDPlusPlus() - Constructor for class org.apache.spark.graphx.lib.SVDPlusPlus
-
- SVDPlusPlus.Conf - Class in org.apache.spark.graphx.lib
-
Configuration parameters for SVDPlusPlus.
- SVMDataGenerator - Class in org.apache.spark.mllib.util
-
Generate sample data used for SVM.
- SVMDataGenerator() - Constructor for class org.apache.spark.mllib.util.SVMDataGenerator
-
- SVMModel - Class in org.apache.spark.mllib.classification
-
Model for Support Vector Machines (SVMs).
- SVMModel(Vector, double) - Constructor for class org.apache.spark.mllib.classification.SVMModel
-
- SVMWithSGD - Class in org.apache.spark.mllib.classification
-
Train a Support Vector Machine (SVM) using Stochastic Gradient Descent.
- SVMWithSGD() - Constructor for class org.apache.spark.mllib.classification.SVMWithSGD
-
Construct a SVM object with default parameters: {stepSize: 1.0, numIterations: 100,
regParam: 0.01, miniBatchFraction: 1.0}.
- symbolToColumn(Symbol) - Method in class org.apache.spark.sql.SQLImplicits
-
An implicit conversion that turns a Scala
Symbol
into a
Column
.
- symlink(File, File) - Static method in class org.apache.spark.util.Utils
-
Creates a symlink.
- symmetricEigs(Function1<DenseVector<Object>, DenseVector<Object>>, int, int, double, int) - Static method in class org.apache.spark.mllib.linalg.EigenValueDecomposition
-
Compute the leading k eigenvalues and eigenvectors on a symmetric square matrix using ARPACK.
- syr(double, Vector, DenseMatrix) - Static method in class org.apache.spark.ml.linalg.BLAS
-
A := alpha * x * x^T^ + A
- syr(double, Vector, DenseMatrix) - Static method in class org.apache.spark.mllib.linalg.BLAS
-
A := alpha * x * x^T^ + A
- SYSTEM_DEFAULT() - Static method in class org.apache.spark.sql.types.DecimalType
-
- SYSTEM_PROPERTIES_FIELD_NUMBER - Static variable in class org.apache.spark.status.protobuf.StoreTypes.ApplicationEnvironmentInfo
-
- systemProperties() - Method in class org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-