public class QueryExecutionErrors
extends Object
QueryCompilationErrors
.Constructor and Description |
---|
QueryExecutionErrors() |
Modifier and Type | Method and Description |
---|---|
static org.apache.spark.SparkUnsupportedOperationException |
addFilesWithAbsolutePathUnsupportedError(String commitProtocol) |
static Throwable |
addNewFunctionMismatchedWithFunctionError(String funcName) |
static RuntimeException |
aesCryptoError(String detailMessage) |
static RuntimeException |
aesModeUnsupportedError(String mode,
String padding) |
static RuntimeException |
aesUnsupportedAad(String mode) |
static RuntimeException |
aesUnsupportedIv(String mode) |
static org.apache.spark.SparkUnsupportedOperationException |
alterTableWithDropPartitionAndPurgeUnsupportedError() |
static org.apache.spark.SparkDateTimeException |
ansiDateTimeError(Exception e) |
static org.apache.spark.SparkDateTimeException |
ansiDateTimeParseError(Exception e) |
static IllegalArgumentException |
ansiIllegalArgumentError(IllegalArgumentException e) |
static org.apache.spark.SparkIllegalArgumentException |
ansiIllegalArgumentError(String message) |
static org.apache.spark.SparkArithmeticException |
arithmeticOverflowError(ArithmeticException e) |
static String |
arithmeticOverflowError$default$2() |
static org.apache.spark.sql.catalyst.trees.SQLQueryContext |
arithmeticOverflowError$default$3() |
static org.apache.spark.SparkUnsupportedOperationException |
attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema) |
static org.apache.spark.SparkFileNotFoundException |
batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile) |
static org.apache.spark.SparkArithmeticException |
binaryArithmeticCauseOverflowError(short eval1,
String symbol,
short eval2) |
static org.apache.spark.SparkUnsupportedOperationException |
buildReaderUnsupportedForFileFormatError(String format) |
static Throwable |
cannotAcquireMemoryToBuildLongHashedRelationError(long size,
long got) |
static Throwable |
cannotAcquireMemoryToBuildUnsafeHashedRelationError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotAllocateMemoryToGrowBytesToBytesMapError() |
static Throwable |
cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes,
long dataSize) |
static Throwable |
cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows,
long numRows) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotBuildHashedRelationLargerThan8GError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotBuildHashedRelationWithUniqueKeysExceededError() |
static Throwable |
cannotCastError(DataType from,
DataType to) |
static Throwable |
cannotCastFromNullTypeError(DataType to) |
static ArithmeticException |
cannotChangeDecimalPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath) |
static Throwable |
cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotCloneOrCopyReadOnlySQLConfError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotCompareCostWithTargetCostError(String cost) |
static Throwable |
cannotConvertCatalystValueToProtobufEnumTypeError(scala.collection.Seq<String> sqlColumn,
String protobufColumn,
String data,
String enumString) |
static Throwable |
cannotConvertOrcTimestampNTZToTimestampLTZError() |
static Throwable |
cannotConvertOrcTimestampToTimestampNTZError() |
static org.apache.spark.SparkRuntimeException |
cannotCreateArrayWithElementsExceedLimitError(long numElements,
String additionalErrorMessage) |
static Throwable |
cannotCreateColumnarReaderError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotCreateJDBCTableWithPartitionsError() |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForDataTypeError(DataType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForDecimalTypeError(DecimalType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForTypeError(DecimalType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateStagingDirError(String message,
java.io.IOException e) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotExecuteStreamingRelationExecError() |
static Throwable |
cannotFetchTablesOfDatabaseError(String dbName,
Exception e) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotFindConstructorForTypeError(String tpe) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotFindEncoderForTypeError(String typeName) |
static Throwable |
cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static org.apache.spark.SparkIllegalArgumentException |
cannotGenerateCodeForIncomparableTypeError(String codeType,
DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotGetEventTimeWatermarkError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotGetJdbcTypeError(DataType dt) |
static org.apache.spark.SparkRuntimeException |
cannotGetOuterPointerForInnerClassError(Class<?> innerCls) |
static org.apache.spark.SparkRuntimeException |
cannotGetSQLConfInSchedulerEventLoopThreadError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotHaveCircularReferencesInBeanClassError(Class<?> clazz) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotHaveCircularReferencesInClassError(String t) |
static Throwable |
cannotInstantiateAbstractCatalogPluginClassError(String name,
String pluginClassName,
Exception e) |
static org.apache.spark.SparkIllegalArgumentException |
cannotInterpolateClassIntoCodeBlockError(Object arg) |
static Throwable |
cannotLoadUserDefinedTypeError(String name,
String userClass) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotMergeClassWithOtherClassError(String className,
String otherClass) |
static Throwable |
cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale,
int rightScale) |
static Throwable |
cannotMergeIncompatibleDataTypesError(DataType left,
DataType right) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotMutateReadOnlySQLConfError() |
static Throwable |
cannotParseDecimalError() |
static org.apache.spark.SparkRuntimeException |
cannotParseJsonArraysAsStructsError(String recordStr) |
static org.apache.spark.SparkRuntimeException |
cannotParseJSONFieldError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken jsonType,
DataType dataType) |
static org.apache.spark.SparkRuntimeException |
cannotParseJSONFieldError(String fieldName,
String fieldValue,
com.fasterxml.jackson.core.JsonToken jsonType,
DataType dataType) |
static org.apache.spark.SparkIllegalArgumentException |
cannotParseStatisticAsPercentileError(String stats,
NumberFormatException e) |
static org.apache.spark.SparkRuntimeException |
cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken token,
DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotPurgeAsBreakInternalStateError() |
static Throwable |
cannotReadFilesError(Throwable e,
String path) |
static Throwable |
cannotReadFooterForFileError(org.apache.hadoop.fs.Path file,
Exception e) |
static Throwable |
cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e,
String fieldType,
String fieldName) |
static org.apache.spark.SparkRuntimeException |
cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotRemoveReservedPropertyError(String property) |
static Throwable |
cannotRestorePermissionsForPathError(org.apache.hadoop.fs.permission.FsPermission permission,
org.apache.hadoop.fs.Path path) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1,
scala.collection.immutable.Map<String,String> props2,
scala.collection.immutable.Set<String> conflictKeys) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSetTimeoutDurationError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSetTimeoutTimestampError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator) |
static org.apache.spark.SparkIllegalArgumentException |
cannotTranslateNonNullValueForFieldError(int pos) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName,
org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) |
static ArithmeticException |
castingCauseOverflowError(Object t,
DataType from,
DataType to) |
static ArithmeticException |
castingCauseOverflowErrorInTableInsert(DataType from,
DataType to,
String columnName) |
static Throwable |
catalogFailToCallPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogFailToFindPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotFoundError(String name) |
static Throwable |
catalogPluginClassNotFoundForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotImplementedError(String name,
String pluginClassName) |
static org.apache.spark.SparkRuntimeException |
classHasUnexpectedSerializerError(String clsName,
org.apache.spark.sql.catalyst.expressions.Expression objSerializer) |
static org.apache.spark.SparkRuntimeException |
classUnsupportedByMapObjectsError(Class<?> cls) |
static org.apache.spark.SparkUnsupportedOperationException |
cleanUpSourceFilesUnsupportedError() |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
commentOnTableUnsupportedError() |
static Throwable |
commitDeniedError(int partId,
long taskId,
int attemptId,
int stageId,
int stageAttempt) |
static Throwable |
comparatorReturnsNull(String firstValue,
String secondValue) |
static Throwable |
compilerError(org.codehaus.commons.compiler.CompileException e) |
static org.apache.spark.SparkRuntimeException |
concatArraysWithElementsExceedLimitError(long numberOfElements) |
static org.apache.spark.SparkConcurrentModificationException |
concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className) |
static Throwable |
concurrentQueryInstanceError() |
static Throwable |
concurrentStreamLogUpdate(long batchId) |
static org.apache.spark.SparkRuntimeException |
constructorNotFoundError(String cls) |
static org.apache.spark.SparkUnsupportedOperationException |
continuousProcessingUnsupportedByDataSourceError(String sourceName) |
static Throwable |
convertHiveTableToCatalogTableError(SparkException e,
String dbName,
String tableName) |
static org.apache.spark.SparkRuntimeException |
createArrayWithElementsExceedLimitError(Object count) |
static org.apache.spark.SparkIllegalArgumentException |
createStreamingSourceNotSpecifySchemaError() |
static org.apache.spark.SparkUnsupportedOperationException |
customCollectionClsNotResolvedError() |
static Throwable |
databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB) |
static org.apache.spark.SparkIllegalArgumentException |
dataPathNotSpecifiedError() |
static org.apache.spark.SparkClassNotFoundException |
dataSourceNotFoundError(String provider,
Throwable error) |
static org.apache.spark.SparkUnsupportedOperationException |
dataTypeUnexpectedError(DataType dataType) |
static Throwable |
dataTypeUnsupportedError(String dataType,
String failure) |
static org.apache.spark.SparkUnsupportedOperationException |
dataTypeUnsupportedYetError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
ddlUnsupportedTemporarilyError(String ddl) |
static org.apache.spark.SparkArithmeticException |
decimalPrecisionExceedsMaxPrecisionError(int precision,
int maxPrecision) |
static org.apache.spark.SparkUnsupportedOperationException |
decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static Throwable |
defaultDatabaseNotExistsError(String defaultDatabase) |
static ArithmeticException |
divideByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkUnsupportedOperationException |
doExecuteBroadcastNotImplementedError(String nodeName) |
static org.apache.spark.SparkUnsupportedOperationException |
dropTableWithPurgeUnsupportedError() |
static org.apache.spark.SparkUnsupportedOperationException |
duplicatedFieldNameInArrowStructError(scala.collection.Seq<String> fieldNames) |
static org.apache.spark.SparkRuntimeException |
duplicateMapKeyFoundError(Object key) |
static Throwable |
dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key) |
static Throwable |
dynamicPartitionOverwriteUnsupportedByTableError(Table table) |
static org.apache.spark.SparkUnsupportedOperationException |
elementsOfTupleExceedLimitError() |
static org.apache.spark.SparkRuntimeException |
emptyJsonFieldValueError(DataType dataType) |
static org.apache.spark.SparkIllegalArgumentException |
emptyOptionError(String optionName) |
static Throwable |
endOfIteratorError() |
static Throwable |
endOfStreamError() |
static org.apache.spark.SparkRuntimeException |
exceedMapSizeLimitError(int size) |
static org.apache.spark.SparkRuntimeException |
exceedMaxLimit(int limit) |
static Throwable |
executeBroadcastTimeoutError(long timeout,
scala.Option<java.util.concurrent.TimeoutException> ex) |
static org.apache.spark.SparkUnsupportedOperationException |
executeCodePathUnsupportedError(String execName) |
static org.apache.spark.SparkRuntimeException |
expressionDecodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static org.apache.spark.SparkRuntimeException |
expressionEncodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static Throwable |
failedExecuteUserDefinedFunctionError(String functionName,
String inputTypes,
String outputType,
Throwable e) |
static Throwable |
failedMergingSchemaError(StructType leftSchema,
StructType rightSchema,
SparkException e) |
static org.apache.spark.SparkRuntimeException |
failedParsingStructTypeError(String raw) |
static Throwable |
failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static org.apache.spark.SparkRuntimeException |
failedToCastValueToDataTypeForPartitionColumnError(String value,
DataType dataType,
String columnName) |
static String |
failedToCompileMsg(Exception e) |
static SparkException |
failedToExecuteQueryError(Throwable e) |
static Throwable |
failedToGenerateEpochMarkerError(Throwable failureReason) |
static Throwable |
failedToInstantiateConstructorForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
failedToMergeIncompatibleSchemasError(StructType left,
StructType right,
Throwable e) |
static Throwable |
failedToPushRowIntoRowQueueError(String rowQueue) |
static Throwable |
failedToReadDataError(Throwable failureReason) |
static Throwable |
failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
int keySize) |
static Throwable |
failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
String message) |
static org.apache.spark.SparkRuntimeException |
failToConvertValueToJsonError(Object value,
Class<?> cls,
DataType dataType) |
static org.apache.spark.SparkUpgradeException |
failToParseDateTimeInNewParserError(String s,
Throwable e) |
static org.apache.spark.SparkUpgradeException |
failToRecognizePatternAfterUpgradeError(String pattern,
Throwable e) |
static org.apache.spark.SparkRuntimeException |
failToRecognizePatternError(String pattern,
Throwable e) |
static org.apache.spark.SparkSecurityException |
failToSetOriginalACLBackError(String aclEntries,
org.apache.hadoop.fs.Path path,
Throwable e) |
static org.apache.spark.SparkIllegalArgumentException |
fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema,
StructType v1Schema) |
static org.apache.spark.SparkRuntimeException |
fieldCannotBeNullError(int index,
String fieldName) |
static String |
fieldCannotBeNullMsg(int index,
String fieldName) |
static org.apache.spark.SparkDateTimeException |
fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field,
int actual,
int expected,
java.time.LocalDate candidate) |
static org.apache.spark.SparkUnsupportedOperationException |
fieldIndexOnRowWithoutSchemaError() |
static Throwable |
fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status,
int maxLength) |
static org.apache.spark.SparkFileNotFoundException |
fileNotFoundError(java.io.FileNotFoundException e) |
static org.apache.spark.SparkRuntimeException |
flattenArraysWithElementsExceedLimitError(long numberOfElements) |
static Throwable |
foreachWriterAbortedDueToTaskFailureError() |
static org.apache.spark.SparkRuntimeException |
foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName,
String matchedOrcFields) |
static org.apache.spark.SparkRuntimeException |
foundDuplicateFieldInFieldIdLookupModeError(int requiredId,
String matchedFields) |
static org.apache.spark.SparkRuntimeException |
foundNullValueForNotNullableFieldError(String name) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
getParentLoggerNotImplementedError(String className) |
static org.apache.spark.SparkRuntimeException |
getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e) |
static QueryContext[] |
getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static String |
getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static org.apache.spark.SparkUnsupportedOperationException |
getTablesByTypeUnsupportedByHiveVersionError() |
static org.apache.spark.SparkUnsupportedOperationException |
hiveTableWithAnsiIntervalsError(org.apache.spark.sql.catalyst.TableIdentifier table) |
static Throwable |
hllInvalidInputSketchBuffer(String function) |
static Throwable |
hllInvalidLgK(String function,
int min,
int max,
String value) |
static Throwable |
hllUnionDifferentLgK(int left,
int right,
String function) |
static Throwable |
illegalLocationClauseForViewPartitionError() |
static Throwable |
illegalUrlError(org.apache.spark.unsafe.types.UTF8String url,
IllegalArgumentException e) |
static Throwable |
incompatibleDataSourceRegisterError(Throwable e) |
static Throwable |
incorrectEndOffset(long rowsPerSecond,
long maxSeconds,
long endSeconds) |
static Throwable |
incorrectRampUpRate(long rowsPerSecond,
long maxSeconds,
long rampUpTimeSeconds) |
static org.apache.spark.SparkUnsupportedOperationException |
inferSchemaUnsupportedForHiveError() |
static org.apache.spark.SparkUnsupportedOperationException |
initialTypeNotTargetDataTypeError(DataType dataType,
String target) |
static org.apache.spark.SparkUnsupportedOperationException |
initialTypeNotTargetDataTypesError(DataType dataType) |
static org.apache.spark.SparkRuntimeException |
inputExternalRowCannotBeNullError() |
static Throwable |
inputFilterNotFullyConvertibleError(String owner) |
static Throwable |
internalCompilerError(org.codehaus.commons.compiler.InternalCompilerException e) |
static ArithmeticException |
intervalArithmeticOverflowError(String message,
String hint,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
intervalDividedByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static RuntimeException |
invalidAesIvLengthError(String mode,
int actualLength) |
static RuntimeException |
invalidAesKeyLengthError(int actualLength) |
static ArrayIndexOutOfBoundsException |
invalidArrayIndexError(int index,
int numElements,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArrayIndexOutOfBoundsException |
invalidBitmapPositionError(long bitPosition,
long bitmapNumBytes) |
static Throwable |
invalidBucketFile(String path) |
static Throwable |
invalidCatalogNameError(String name) |
static ArrayIndexOutOfBoundsException |
invalidElementAtIndexError(int index,
int numElements,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidEmptyLocationError(String location) |
static java.time.DateTimeException |
invalidFractionOfSecondError() |
static RuntimeException |
invalidIndexOfZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkDateTimeException |
invalidInputInCastToDatetimeError(double value,
DataType to,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkDateTimeException |
invalidInputInCastToDatetimeError(org.apache.spark.unsafe.types.UTF8String value,
DataType to,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkNumberFormatException |
invalidInputInCastToNumberError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidInputInConversionError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.unsafe.types.UTF8String fmt,
String hint) |
static org.apache.spark.SparkRuntimeException |
invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidJdbcNumPartitionsError(int n,
String jdbcNumPartitions) |
static org.apache.spark.SparkIllegalArgumentException |
invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel,
String value) |
static Throwable |
invalidKerberosConfigForHiveServer2Error() |
static org.apache.spark.SparkIllegalArgumentException |
invalidNamespaceNameError(String[] namespace) |
static org.apache.spark.SparkIllegalArgumentException |
invalidNumberFormatError(DataType dataType,
String input,
String format) |
static org.apache.spark.SparkUnsupportedOperationException |
invalidPartitionFilterError() |
static RuntimeException |
invalidPatternError(String funcName,
String pattern,
Throwable cause) |
static RuntimeException |
invalidRegexGroupIndexError(String funcName,
int groupCount,
int groupIndex) |
static org.apache.spark.SparkArrayIndexOutOfBoundsException |
invalidStartIndexError(int numRows,
int startIndex) |
static org.apache.spark.SparkUnsupportedOperationException |
invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode) |
static org.apache.spark.SparkIllegalArgumentException |
invalidUrlError(org.apache.spark.unsafe.types.UTF8String url,
java.net.URISyntaxException e) |
static org.apache.spark.SparkUnsupportedOperationException |
latestOffsetNotCalledError() |
static Throwable |
legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath,
String legacyCheckpointDir) |
static Throwable |
legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath,
org.apache.hadoop.fs.Path legacyMetadataPath) |
static RuntimeException |
literalTypeUnsupportedError(Object v) |
static org.apache.spark.SparkClassNotFoundException |
loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf,
scala.collection.Seq<java.net.URL> execJars,
String key,
java.lang.reflect.InvocationTargetException e) |
static Throwable |
locationAlreadyExists(org.apache.spark.sql.catalyst.TableIdentifier tableId,
org.apache.hadoop.fs.Path location) |
static org.apache.spark.SparkRuntimeException |
malformedCSVRecordError(String badRecord) |
static Throwable |
malformedJSONError() |
static Throwable |
malformedProtobufMessageDetectedInMessageParsingError(Throwable e) |
static Throwable |
malformedRecordsDetectedInRecordParsingError(String badRecord,
org.apache.spark.sql.catalyst.util.BadRecordException e) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(DataType dataType) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(Throwable e) |
static org.apache.spark.SparkRuntimeException |
mapDataKeyArrayLengthDiffersFromValueArrayLengthError() |
static org.apache.spark.SparkRuntimeException |
mapSizeExceedArraySizeWhenZipMapError(int size) |
static org.apache.spark.SparkRuntimeException |
mergeCardinalityViolationError() |
static Throwable |
mergeUnsupportedByWindowFunctionError(String funcName) |
static org.apache.spark.SparkRuntimeException |
methodCalledInAnalyzerNotAllowedError() |
static Throwable |
methodNotDeclaredError(String name) |
static Throwable |
methodNotFoundError(Class<?> cls,
String functionName,
scala.collection.Seq<Class<?>> argClasses) |
static org.apache.spark.SparkUnsupportedOperationException |
methodNotImplementedError(String methodName) |
static org.apache.spark.SparkUnsupportedOperationException |
microBatchUnsupportedByDataSourceError(String srcName,
String disabledSources,
Table table) |
static org.apache.spark.SparkIllegalArgumentException |
missingDatabaseLocationError() |
static org.apache.spark.SparkIllegalArgumentException |
missingJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
multiActionAlterError(String tableName) |
static Throwable |
multiFailuresInStageMaterializationError(Throwable error) |
static org.apache.spark.SparkIllegalArgumentException |
multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths) |
static Throwable |
multipleRowSubqueryError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkConcurrentModificationException |
multiStreamingQueriesUsingPathConcurrentlyError(String path,
org.apache.hadoop.fs.FileAlreadyExistsException e) |
static org.apache.spark.SparkIllegalArgumentException |
negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression) |
static org.apache.spark.SparkIllegalArgumentException |
nestedArraysUnsupportedError() |
static org.apache.spark.SparkUnsupportedOperationException |
nestedFieldUnsupportedError(String colName) |
static SparkException |
noDefaultForDataTypeError(DataType dataType) |
static AnalysisException |
nonTimeWindowNotSupportedInStreamingError(scala.collection.Seq<String> windowFuncList,
scala.collection.Seq<String> columnNameList,
scala.collection.Seq<String> windowSpecList,
org.apache.spark.sql.catalyst.trees.Origin origin) |
static Throwable |
noRecordsFromEmptyDataReaderError() |
static Throwable |
noSuchElementExceptionError() |
static Throwable |
notADatasourceRDDPartitionError(Partition split) |
static Throwable |
notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe,
scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables) |
static org.apache.spark.SparkRuntimeException |
notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr) |
static org.apache.spark.SparkRuntimeException |
notOverrideExpectedMethodsError(String className,
String m1,
String m2) |
static org.apache.spark.SparkUnsupportedOperationException |
notPublicClassError(String name) |
static org.apache.spark.SparkRuntimeException |
notSupportNonPrimitiveTypeError() |
static Throwable |
notSupportTypeError(DataType dataType) |
static Throwable |
notUserDefinedTypeError(String name,
String userClass) |
static org.apache.spark.SparkRuntimeException |
nullAsMapKeyNotAllowedError() |
static org.apache.spark.SparkUnsupportedOperationException |
nullLiteralsCannotBeCastedError(String name) |
static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> |
onceStrategyIdempotenceIsBrokenForBatchError(String batchName,
TreeType plan,
TreeType reOptimized) |
static Throwable |
onlySupportDataSourcesProvidingFileFormatError(String providingClass) |
static org.apache.spark.SparkIllegalArgumentException |
orderedOperationUnsupportedByDataTypeError(DataType dataType) |
static org.apache.spark.SparkIllegalArgumentException |
orderedOperationUnsupportedByDataTypeError(String dataType) |
static org.apache.spark.SparkArithmeticException |
outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str) |
static Throwable |
outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName,
DataType dt) |
static ArithmeticException |
overflowInConvError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
overflowInIntegralDivideError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
overflowInSumOfDecimalError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
overwriteTableByUnsupportedExpressionError(Table table) |
static org.apache.spark.SparkUnsupportedOperationException |
pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1,
org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2,
String function) |
static org.apache.spark.SparkRuntimeException |
paramExceedOneCharError(String paramName) |
static Throwable |
paramIsNotBooleanValueError(String paramName) |
static org.apache.spark.SparkRuntimeException |
paramIsNotIntegerError(String paramName,
String value) |
static Throwable |
parentSparkUIToAttachTabNotFoundError() |
static org.apache.spark.SparkRuntimeException |
partitionColumnNotFoundInSchemaError(String col,
StructType schema) |
static RuntimeException |
pivotColumnUnsupportedError(Object v,
DataType dataType) |
static Throwable |
pivotNotAfterGroupByUnsupportedError() |
static org.apache.spark.SparkRuntimeException |
primaryConstructorNotFoundError(Class<?> cls) |
static org.apache.spark.SparkUnsupportedOperationException |
primitiveTypesNotSupportedError() |
static org.apache.spark.SparkFileNotFoundException |
readCurrentFileNotFoundError(java.io.FileNotFoundException e) |
static Throwable |
registeringStreamingQueryListenerError(Exception e) |
static org.apache.spark.SparkRuntimeException |
remoteOperationsUnsupportedError() |
static org.apache.spark.SparkClassNotFoundException |
removedClassInSpark2Error(String className,
Throwable e) |
static org.apache.spark.SparkFileAlreadyExistsException |
renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
renameColumnUnsupportedForOlderMySQLError() |
static Throwable |
renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static Throwable |
renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath) |
static Throwable |
repeatedPivotsUnsupportedError(String clause,
String operation) |
static Throwable |
requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table,
scala.collection.immutable.Map<String,scala.Option<String>> partition) |
static org.apache.spark.SparkRuntimeException |
resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static org.apache.spark.SparkRuntimeException |
rootConverterReturnNullError() |
static org.apache.spark.SparkUnsupportedOperationException |
rowLargerThan256MUnsupportedError() |
static Throwable |
ruleIdNotFoundForRuleError(String ruleName) |
static Throwable |
saveModeUnsupportedError(Object saveMode,
boolean pathExists) |
static org.apache.spark.SparkClassNotFoundException |
serDeInterfaceNotFoundError(NoClassDefFoundError e) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInReadingDatesError(String format,
String config,
String option) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInWritingDatesError(String format,
String config) |
static org.apache.spark.SparkNoSuchElementException |
sqlConfigNotFoundError(String key) |
static Throwable |
stateNotDefinedOrAlreadyRemovedError() |
static org.apache.spark.SparkIllegalArgumentException |
statisticNotRecognizedError(String stats) |
static org.apache.spark.SparkUnsupportedOperationException |
streamedOperatorUnsupportedByDataSourceError(String className,
String operator) |
static Throwable |
subprocessExitedError(int exitCode,
org.apache.spark.util.CircularBuffer stderrBuffer,
Throwable cause) |
static Throwable |
taskFailedWhileWritingRowsError(String path,
Throwable cause) |
static ArithmeticException |
timestampAddOverflowError(long micros,
int amount,
String unit) |
static String |
toDSOption(String option) |
static org.apache.spark.SparkIllegalArgumentException |
tooManyArrayElementsError(int numElements,
int elementSize) |
static String |
toSQLConf(String conf) |
static String |
toSQLConfVal(String conf) |
static String |
toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e) |
static String |
toSQLId(scala.collection.Seq<String> parts) |
static String |
toSQLId(String parts) |
static String |
toSQLSchema(String schema) |
static String |
toSQLStmt(String text) |
static String |
toSQLType(org.apache.spark.sql.types.AbstractDataType t) |
static String |
toSQLType(String text) |
static String |
toSQLValue(double value) |
static String |
toSQLValue(float value) |
static String |
toSQLValue(int value) |
static String |
toSQLValue(long value) |
static String |
toSQLValue(Object v,
DataType t) |
static String |
toSQLValue(short value) |
static String |
toSQLValue(String value) |
static String |
toSQLValue(org.apache.spark.unsafe.types.UTF8String value) |
static Throwable |
transformationsAndActionsNotInvokedByDriverError() |
static org.apache.spark.SparkUnsupportedOperationException |
truncateMultiPartitionUnsupportedError(String tableName) |
static org.apache.spark.SparkIllegalArgumentException |
typeUnsupportedError(DataType dataType) |
static Throwable |
unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToCreateTableAsFailedToCreateDirectoryError(String table,
org.apache.hadoop.fs.Path defaultTableLocation,
java.io.IOException e) |
static Throwable |
unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToDropTableAsFailedToDeleteDirectoryError(String table,
org.apache.hadoop.fs.Path dir,
java.io.IOException e) |
static Throwable |
unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath,
java.io.IOException e) |
static Throwable |
unableToRenameTableAsFailedToRenameDirectoryError(String oldName,
String newName,
org.apache.hadoop.fs.Path oldDir,
java.io.IOException e) |
static org.apache.spark.SparkArithmeticException |
unaryMinusCauseOverflowError(int originValue) |
static org.apache.spark.SparkIllegalArgumentException |
unexpectedAccumulableUpdateValueError(Object o) |
static org.apache.spark.SparkRuntimeException |
unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op,
String pos) |
static org.apache.spark.SparkRuntimeException |
unexpectedValueForLengthInFunctionError(String prettyName) |
static org.apache.spark.SparkRuntimeException |
unexpectedValueForStartInFunctionError(String prettyName) |
static org.apache.spark.SparkRuntimeException |
unexpectedWindowFunctionFrameError(String frame) |
static org.apache.spark.SparkRuntimeException |
unionArrayWithElementsExceedLimitError(int length) |
static org.apache.spark.SparkIllegalArgumentException |
unknownColumnError(String unknownColumn) |
static org.apache.spark.SparkRuntimeException |
unreachableError(String err) |
static String |
unreachableError$default$1() |
static org.apache.spark.SparkUnsupportedOperationException |
unrecognizedCompressionSchemaTypeIDError(int typeId) |
static Throwable |
unrecognizedSqlTypeError(String jdbcTypeId,
String typeName) |
static ArithmeticException |
unscaledValueTooLargeForPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
unsupportedArrayElementTypeBasedOnBinaryError(DataType dt) |
static org.apache.spark.SparkRuntimeException |
unsupportedArrayTypeError(Class<?> clazz) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedArrowTypeError(org.apache.arrow.vector.types.pojo.ArrowType typeName) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedCommentNamespaceError(String namespace) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedDataTypeError(DataType typeName) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedDropNamespaceError(String namespace) |
static org.apache.spark.SparkRuntimeException |
unsupportedEncoderError() |
static org.apache.spark.SparkRuntimeException |
unsupportedFieldNameError(String fieldName) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedHiveMetastoreVersionError(String version,
String key) |
static org.apache.spark.SparkRuntimeException |
unsupportedJavaTypeError(Class<?> clazz) |
static org.apache.spark.SparkSQLException |
unsupportedJdbcTypeError(String content) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedMultipleBucketTransformsError() |
static SparkException |
unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperandTypeForSizeFunctionError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperationExceptionError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperationForDataTypeError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedPartitionTransformError(Transform transform) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedPurgePartitionError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedPurgeTableError() |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedRemoveNamespaceCommentError(String namespace) |
static SparkException |
unsupportedRoundingMode(scala.Enumeration.Value roundMode) |
static Throwable |
unsupportedSchemaColumnConvertError(String filePath,
String column,
String logicalType,
String physicalType,
Exception e) |
static Throwable |
unsupportedTableChangeError(IllegalArgumentException e) |
static Throwable |
unsupportedTypeError(DataType dataType) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedUpdateColumnNullabilityError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedUserSpecifiedSchemaError() |
static Throwable |
useDictionaryEncodingWhenDictionaryOverflowError() |
static Throwable |
userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt) |
static org.apache.spark.SparkUnsupportedOperationException |
userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider) |
static Throwable |
valueIsNullError(int index) |
static Throwable |
writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts,
int maxDynamicPartitions,
String maxDynamicPartitionsKey) |
static org.apache.spark.SparkUnsupportedOperationException |
writeUnsupportedForBinaryFileDataSourceError() |
static Throwable |
writingJobFailedError(Throwable cause) |
public static Throwable cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator)
public static ArithmeticException castingCauseOverflowError(Object t, DataType from, DataType to)
public static ArithmeticException castingCauseOverflowErrorInTableInsert(DataType from, DataType to, String columnName)
public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkIllegalArgumentException invalidInputInConversionError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.unsafe.types.UTF8String fmt, String hint)
public static Throwable cannotCastFromNullTypeError(DataType to)
public static Throwable cannotParseDecimalError()
public static Throwable dataTypeUnsupportedError(String dataType, String failure)
public static Throwable failedExecuteUserDefinedFunctionError(String functionName, String inputTypes, String outputType, Throwable e)
public static ArithmeticException divideByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException intervalDividedByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArrayIndexOutOfBoundsException invalidArrayIndexError(int index, int numElements, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArrayIndexOutOfBoundsException invalidElementAtIndexError(int index, int numElements, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArrayIndexOutOfBoundsException invalidBitmapPositionError(long bitPosition, long bitmapNumBytes)
public static java.time.DateTimeException invalidFractionOfSecondError()
public static org.apache.spark.SparkDateTimeException ansiDateTimeParseError(Exception e)
public static org.apache.spark.SparkDateTimeException ansiDateTimeError(Exception e)
public static org.apache.spark.SparkIllegalArgumentException ansiIllegalArgumentError(String message)
public static IllegalArgumentException ansiIllegalArgumentError(IllegalArgumentException e)
public static ArithmeticException overflowInSumOfDecimalError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException overflowInIntegralDivideError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException overflowInConvError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException mapSizeExceedArraySizeWhenZipMapError(int size)
public static RuntimeException literalTypeUnsupportedError(Object v)
public static RuntimeException pivotColumnUnsupportedError(Object v, DataType dataType)
public static SparkException noDefaultForDataTypeError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException orderedOperationUnsupportedByDataTypeError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException orderedOperationUnsupportedByDataTypeError(String dataType)
public static RuntimeException invalidRegexGroupIndexError(String funcName, int groupCount, int groupIndex)
public static org.apache.spark.SparkIllegalArgumentException invalidUrlError(org.apache.spark.unsafe.types.UTF8String url, java.net.URISyntaxException e)
public static Throwable illegalUrlError(org.apache.spark.unsafe.types.UTF8String url, IllegalArgumentException e)
public static Throwable mergeUnsupportedByWindowFunctionError(String funcName)
public static org.apache.spark.SparkUnsupportedOperationException dataTypeUnexpectedError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException typeUnsupportedError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression)
public static Throwable addNewFunctionMismatchedWithFunctionError(String funcName)
public static org.apache.spark.SparkIllegalArgumentException cannotGenerateCodeForIncomparableTypeError(String codeType, DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException cannotInterpolateClassIntoCodeBlockError(Object arg)
public static org.apache.spark.SparkUnsupportedOperationException customCollectionClsNotResolvedError()
public static org.apache.spark.SparkRuntimeException classUnsupportedByMapObjectsError(Class<?> cls)
public static org.apache.spark.SparkRuntimeException nullAsMapKeyNotAllowedError()
public static Throwable methodNotDeclaredError(String name)
public static Throwable methodNotFoundError(Class<?> cls, String functionName, scala.collection.Seq<Class<?>> argClasses)
public static org.apache.spark.SparkRuntimeException constructorNotFoundError(String cls)
public static SparkException unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType)
public static org.apache.spark.SparkRuntimeException notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr)
public static org.apache.spark.SparkRuntimeException unsupportedEncoderError()
public static org.apache.spark.SparkRuntimeException notOverrideExpectedMethodsError(String className, String m1, String m2)
public static org.apache.spark.SparkRuntimeException failToConvertValueToJsonError(Object value, Class<?> cls, DataType dataType)
public static org.apache.spark.SparkRuntimeException unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op, String pos)
public static SparkException unsupportedRoundingMode(scala.Enumeration.Value roundMode)
public static org.apache.spark.SparkRuntimeException resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static org.apache.spark.SparkRuntimeException inputExternalRowCannotBeNullError()
public static String fieldCannotBeNullMsg(int index, String fieldName)
public static org.apache.spark.SparkRuntimeException fieldCannotBeNullError(int index, String fieldName)
public static Throwable unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToCreateTableAsFailedToCreateDirectoryError(String table, org.apache.hadoop.fs.Path defaultTableLocation, java.io.IOException e)
public static Throwable unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToDropTableAsFailedToDeleteDirectoryError(String table, org.apache.hadoop.fs.Path dir, java.io.IOException e)
public static Throwable unableToRenameTableAsFailedToRenameDirectoryError(String oldName, String newName, org.apache.hadoop.fs.Path oldDir, java.io.IOException e)
public static Throwable unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath, java.io.IOException e)
public static org.apache.spark.SparkUnsupportedOperationException methodNotImplementedError(String methodName)
public static org.apache.spark.SparkArithmeticException arithmeticOverflowError(ArithmeticException e)
public static org.apache.spark.SparkArithmeticException unaryMinusCauseOverflowError(int originValue)
public static org.apache.spark.SparkArithmeticException binaryArithmeticCauseOverflowError(short eval1, String symbol, short eval2)
public static ArithmeticException intervalArithmeticOverflowError(String message, String hint, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static String failedToCompileMsg(Exception e)
public static Throwable internalCompilerError(org.codehaus.commons.compiler.InternalCompilerException e)
public static Throwable compilerError(org.codehaus.commons.compiler.CompileException e)
public static Throwable unsupportedTableChangeError(IllegalArgumentException e)
public static Throwable notADatasourceRDDPartitionError(Partition split)
public static org.apache.spark.SparkIllegalArgumentException dataPathNotSpecifiedError()
public static org.apache.spark.SparkIllegalArgumentException createStreamingSourceNotSpecifySchemaError()
public static org.apache.spark.SparkUnsupportedOperationException streamedOperatorUnsupportedByDataSourceError(String className, String operator)
public static AnalysisException nonTimeWindowNotSupportedInStreamingError(scala.collection.Seq<String> windowFuncList, scala.collection.Seq<String> columnNameList, scala.collection.Seq<String> windowSpecList, org.apache.spark.sql.catalyst.trees.Origin origin)
public static org.apache.spark.SparkIllegalArgumentException multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths)
public static org.apache.spark.SparkClassNotFoundException dataSourceNotFoundError(String provider, Throwable error)
public static org.apache.spark.SparkClassNotFoundException removedClassInSpark2Error(String className, Throwable e)
public static Throwable incompatibleDataSourceRegisterError(Throwable e)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInReadingDatesError(String format, String config, String option)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInWritingDatesError(String format, String config)
public static org.apache.spark.SparkUnsupportedOperationException buildReaderUnsupportedForFileFormatError(String format)
public static Throwable taskFailedWhileWritingRowsError(String path, Throwable cause)
public static org.apache.spark.SparkFileNotFoundException readCurrentFileNotFoundError(java.io.FileNotFoundException e)
public static Throwable saveModeUnsupportedError(Object saveMode, boolean pathExists)
public static Throwable cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath)
public static Throwable cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path)
public static org.apache.spark.SparkRuntimeException failedToCastValueToDataTypeForPartitionColumnError(String value, DataType dataType, String columnName)
public static Throwable endOfStreamError()
public static org.apache.spark.SparkIllegalArgumentException fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema, StructType v1Schema)
public static Throwable noRecordsFromEmptyDataReaderError()
public static org.apache.spark.SparkFileNotFoundException fileNotFoundError(java.io.FileNotFoundException e)
public static Throwable unsupportedSchemaColumnConvertError(String filePath, String column, String logicalType, String physicalType, Exception e)
public static Throwable cannotReadFilesError(Throwable e, String path)
public static Throwable cannotCreateColumnarReaderError()
public static org.apache.spark.SparkIllegalArgumentException invalidNamespaceNameError(String[] namespace)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPartitionTransformError(Transform transform)
public static org.apache.spark.SparkIllegalArgumentException missingDatabaseLocationError()
public static org.apache.spark.SparkUnsupportedOperationException cannotRemoveReservedPropertyError(String property)
public static Throwable writingJobFailedError(Throwable cause)
public static Throwable commitDeniedError(int partId, long taskId, int attemptId, int stageId, int stageAttempt)
public static org.apache.spark.SparkUnsupportedOperationException cannotCreateJDBCTableWithPartitionsError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedUserSpecifiedSchemaError()
public static org.apache.spark.SparkUnsupportedOperationException writeUnsupportedForBinaryFileDataSourceError()
public static Throwable fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status, int maxLength)
public static org.apache.spark.SparkRuntimeException unsupportedFieldNameError(String fieldName)
public static org.apache.spark.SparkIllegalArgumentException cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static org.apache.spark.SparkIllegalArgumentException missingJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static org.apache.spark.SparkIllegalArgumentException emptyOptionError(String optionName)
public static org.apache.spark.SparkIllegalArgumentException invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel, String value)
public static org.apache.spark.SparkIllegalArgumentException cannotGetJdbcTypeError(DataType dt)
public static Throwable unrecognizedSqlTypeError(String jdbcTypeId, String typeName)
public static org.apache.spark.SparkSQLException unsupportedJdbcTypeError(String content)
public static org.apache.spark.SparkIllegalArgumentException unsupportedArrayElementTypeBasedOnBinaryError(DataType dt)
public static org.apache.spark.SparkIllegalArgumentException nestedArraysUnsupportedError()
public static org.apache.spark.SparkIllegalArgumentException cannotTranslateNonNullValueForFieldError(int pos)
public static org.apache.spark.SparkIllegalArgumentException invalidJdbcNumPartitionsError(int n, String jdbcNumPartitions)
public static Throwable multiActionAlterError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException dataTypeUnsupportedYetError(DataType dataType)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationForDataTypeError(DataType dataType)
public static Throwable inputFilterNotFullyConvertibleError(String owner)
public static Throwable cannotReadFooterForFileError(org.apache.hadoop.fs.Path file, Exception e)
public static org.apache.spark.SparkRuntimeException foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName, String matchedOrcFields)
public static org.apache.spark.SparkRuntimeException foundDuplicateFieldInFieldIdLookupModeError(int requiredId, String matchedFields)
public static Throwable failedToMergeIncompatibleSchemasError(StructType left, StructType right, Throwable e)
public static org.apache.spark.SparkUnsupportedOperationException ddlUnsupportedTemporarilyError(String ddl)
public static Throwable executeBroadcastTimeoutError(long timeout, scala.Option<java.util.concurrent.TimeoutException> ex)
public static org.apache.spark.SparkIllegalArgumentException cannotCompareCostWithTargetCostError(String cost)
public static Throwable notSupportTypeError(DataType dataType)
public static org.apache.spark.SparkRuntimeException notSupportNonPrimitiveTypeError()
public static Throwable unsupportedTypeError(DataType dataType)
public static Throwable useDictionaryEncodingWhenDictionaryOverflowError()
public static Throwable endOfIteratorError()
public static Throwable cannotAllocateMemoryToGrowBytesToBytesMapError()
public static Throwable cannotAcquireMemoryToBuildLongHashedRelationError(long size, long got)
public static Throwable cannotAcquireMemoryToBuildUnsafeHashedRelationError()
public static org.apache.spark.SparkUnsupportedOperationException rowLargerThan256MUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationWithUniqueKeysExceededError()
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationLargerThan8GError()
public static Throwable failedToPushRowIntoRowQueueError(String rowQueue)
public static org.apache.spark.SparkRuntimeException unexpectedWindowFunctionFrameError(String frame)
public static org.apache.spark.SparkIllegalArgumentException cannotParseStatisticAsPercentileError(String stats, NumberFormatException e)
public static org.apache.spark.SparkIllegalArgumentException statisticNotRecognizedError(String stats)
public static org.apache.spark.SparkIllegalArgumentException unknownColumnError(String unknownColumn)
public static org.apache.spark.SparkIllegalArgumentException unexpectedAccumulableUpdateValueError(Object o)
public static ArithmeticException unscaledValueTooLargeForPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkArithmeticException decimalPrecisionExceedsMaxPrecisionError(int precision, int maxPrecision)
public static org.apache.spark.SparkArithmeticException outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str)
public static org.apache.spark.SparkRuntimeException unsupportedArrayTypeError(Class<?> clazz)
public static org.apache.spark.SparkRuntimeException unsupportedJavaTypeError(Class<?> clazz)
public static org.apache.spark.SparkRuntimeException failedParsingStructTypeError(String raw)
public static Throwable cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale, int rightScale)
public static Throwable cannotMergeIncompatibleDataTypesError(DataType left, DataType right)
public static org.apache.spark.SparkRuntimeException exceedMapSizeLimitError(int size)
public static org.apache.spark.SparkRuntimeException duplicateMapKeyFoundError(Object key)
public static org.apache.spark.SparkRuntimeException mapDataKeyArrayLengthDiffersFromValueArrayLengthError()
public static Throwable registeringStreamingQueryListenerError(Exception e)
public static Throwable concurrentQueryInstanceError()
public static Throwable concurrentStreamLogUpdate(long batchId)
public static org.apache.spark.SparkRuntimeException cannotParseJsonArraysAsStructsError(String recordStr)
public static org.apache.spark.SparkRuntimeException cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken token, DataType dataType)
public static org.apache.spark.SparkRuntimeException emptyJsonFieldValueError(DataType dataType)
public static org.apache.spark.SparkRuntimeException cannotParseJSONFieldError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType)
public static org.apache.spark.SparkRuntimeException cannotParseJSONFieldError(String fieldName, String fieldValue, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType)
public static org.apache.spark.SparkRuntimeException rootConverterReturnNullError()
public static org.apache.spark.SparkUnsupportedOperationException attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema)
public static org.apache.spark.SparkRuntimeException paramExceedOneCharError(String paramName)
public static org.apache.spark.SparkRuntimeException paramIsNotIntegerError(String paramName, String value)
public static Throwable paramIsNotBooleanValueError(String paramName)
public static org.apache.spark.SparkRuntimeException foundNullValueForNotNullableFieldError(String name)
public static org.apache.spark.SparkRuntimeException malformedCSVRecordError(String badRecord)
public static org.apache.spark.SparkUnsupportedOperationException elementsOfTupleExceedLimitError()
public static org.apache.spark.SparkRuntimeException expressionDecodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static org.apache.spark.SparkRuntimeException expressionEncodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static org.apache.spark.SparkRuntimeException classHasUnexpectedSerializerError(String clsName, org.apache.spark.sql.catalyst.expressions.Expression objSerializer)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperandTypeForSizeFunctionError(DataType dataType)
public static org.apache.spark.SparkRuntimeException unexpectedValueForStartInFunctionError(String prettyName)
public static org.apache.spark.SparkRuntimeException unexpectedValueForLengthInFunctionError(String prettyName)
public static RuntimeException invalidIndexOfZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException concatArraysWithElementsExceedLimitError(long numberOfElements)
public static org.apache.spark.SparkRuntimeException flattenArraysWithElementsExceedLimitError(long numberOfElements)
public static org.apache.spark.SparkRuntimeException createArrayWithElementsExceedLimitError(Object count)
public static org.apache.spark.SparkRuntimeException unionArrayWithElementsExceedLimitError(int length)
public static org.apache.spark.SparkUnsupportedOperationException initialTypeNotTargetDataTypeError(DataType dataType, String target)
public static org.apache.spark.SparkUnsupportedOperationException initialTypeNotTargetDataTypesError(DataType dataType)
public static Throwable malformedRecordsDetectedInSchemaInferenceError(Throwable e)
public static Throwable malformedJSONError()
public static Throwable malformedRecordsDetectedInSchemaInferenceError(DataType dataType)
public static org.apache.spark.SparkUnsupportedOperationException decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static org.apache.spark.SparkRuntimeException methodCalledInAnalyzerNotAllowedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1, scala.collection.immutable.Map<String,String> props2, scala.collection.immutable.Set<String> conflictKeys)
public static org.apache.spark.SparkUnsupportedOperationException pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1, org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2, String function)
public static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> org.apache.spark.SparkRuntimeException onceStrategyIdempotenceIsBrokenForBatchError(String batchName, TreeType plan, TreeType reOptimized)
public static Throwable ruleIdNotFoundForRuleError(String ruleName)
public static org.apache.spark.SparkRuntimeException cannotCreateArrayWithElementsExceedLimitError(long numElements, String additionalErrorMessage)
public static Throwable malformedRecordsDetectedInRecordParsingError(String badRecord, org.apache.spark.sql.catalyst.util.BadRecordException e)
public static org.apache.spark.SparkRuntimeException remoteOperationsUnsupportedError()
public static Throwable invalidKerberosConfigForHiveServer2Error()
public static Throwable parentSparkUIToAttachTabNotFoundError()
public static org.apache.spark.SparkUnsupportedOperationException inferSchemaUnsupportedForHiveError()
public static Throwable requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table, scala.collection.immutable.Map<String,scala.Option<String>> partition)
public static Throwable dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key)
public static org.apache.spark.SparkRuntimeException cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath)
public static org.apache.spark.SparkRuntimeException cannotCreateStagingDirError(String message, java.io.IOException e)
public static org.apache.spark.SparkClassNotFoundException serDeInterfaceNotFoundError(NoClassDefFoundError e)
public static Throwable convertHiveTableToCatalogTableError(SparkException e, String dbName, String tableName)
public static Throwable cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e, String fieldType, String fieldName)
public static org.apache.spark.SparkUnsupportedOperationException getTablesByTypeUnsupportedByHiveVersionError()
public static org.apache.spark.SparkUnsupportedOperationException dropTableWithPurgeUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException alterTableWithDropPartitionAndPurgeUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException invalidPartitionFilterError()
public static org.apache.spark.SparkRuntimeException getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedHiveMetastoreVersionError(String version, String key)
public static org.apache.spark.SparkClassNotFoundException loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf, scala.collection.Seq<java.net.URL> execJars, String key, java.lang.reflect.InvocationTargetException e)
public static Throwable cannotFetchTablesOfDatabaseError(String dbName, Exception e)
public static Throwable illegalLocationClauseForViewPartitionError()
public static Throwable renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static org.apache.spark.SparkFileAlreadyExistsException renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath)
public static Throwable renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath)
public static Throwable failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static Throwable legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath, org.apache.hadoop.fs.Path legacyMetadataPath)
public static org.apache.spark.SparkRuntimeException partitionColumnNotFoundInSchemaError(String col, StructType schema)
public static Throwable stateNotDefinedOrAlreadyRemovedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutDurationError()
public static org.apache.spark.SparkUnsupportedOperationException cannotGetEventTimeWatermarkError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutTimestampError()
public static org.apache.spark.SparkFileNotFoundException batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile)
public static org.apache.spark.SparkConcurrentModificationException multiStreamingQueriesUsingPathConcurrentlyError(String path, org.apache.hadoop.fs.FileAlreadyExistsException e)
public static org.apache.spark.SparkUnsupportedOperationException addFilesWithAbsolutePathUnsupportedError(String commitProtocol)
public static org.apache.spark.SparkUnsupportedOperationException microBatchUnsupportedByDataSourceError(String srcName, String disabledSources, Table table)
public static org.apache.spark.SparkUnsupportedOperationException cannotExecuteStreamingRelationExecError()
public static org.apache.spark.SparkUnsupportedOperationException invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode)
public static Throwable invalidCatalogNameError(String name)
public static Throwable catalogPluginClassNotFoundError(String name)
public static Throwable catalogPluginClassNotImplementedError(String name, String pluginClassName)
public static Throwable catalogPluginClassNotFoundForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToFindPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToCallPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable cannotInstantiateAbstractCatalogPluginClassError(String name, String pluginClassName, Exception e)
public static Throwable failedToInstantiateConstructorForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable noSuchElementExceptionError()
public static org.apache.spark.SparkNoSuchElementException sqlConfigNotFoundError(String key)
public static org.apache.spark.SparkUnsupportedOperationException cannotMutateReadOnlySQLConfError()
public static org.apache.spark.SparkUnsupportedOperationException cannotCloneOrCopyReadOnlySQLConfError()
public static org.apache.spark.SparkRuntimeException cannotGetSQLConfInSchedulerEventLoopThreadError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationExceptionError()
public static org.apache.spark.SparkUnsupportedOperationException nullLiteralsCannotBeCastedError(String name)
public static Throwable notUserDefinedTypeError(String name, String userClass)
public static Throwable cannotLoadUserDefinedTypeError(String name, String userClass)
public static org.apache.spark.SparkUnsupportedOperationException notPublicClassError(String name)
public static org.apache.spark.SparkUnsupportedOperationException primitiveTypesNotSupportedError()
public static org.apache.spark.SparkUnsupportedOperationException fieldIndexOnRowWithoutSchemaError()
public static Throwable valueIsNullError(int index)
public static Throwable onlySupportDataSourcesProvidingFileFormatError(String providingClass)
public static Throwable cannotRestorePermissionsForPathError(org.apache.hadoop.fs.permission.FsPermission permission, org.apache.hadoop.fs.Path path)
public static org.apache.spark.SparkSecurityException failToSetOriginalACLBackError(String aclEntries, org.apache.hadoop.fs.Path path, Throwable e)
public static Throwable multiFailuresInStageMaterializationError(Throwable error)
public static org.apache.spark.SparkUnsupportedOperationException unrecognizedCompressionSchemaTypeIDError(int typeId)
public static org.apache.spark.SparkSQLFeatureNotSupportedException getParentLoggerNotImplementedError(String className)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForTypeError(DecimalType t, String parquetType)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForDecimalTypeError(DecimalType t, String parquetType)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForDataTypeError(DataType t, String parquetType)
public static org.apache.spark.SparkUnsupportedOperationException cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider)
public static org.apache.spark.SparkUnsupportedOperationException cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException truncateMultiPartitionUnsupportedError(String tableName)
public static Throwable overwriteTableByUnsupportedExpressionError(Table table)
public static Throwable dynamicPartitionOverwriteUnsupportedByTableError(Table table)
public static Throwable failedMergingSchemaError(StructType leftSchema, StructType rightSchema, SparkException e)
public static Throwable cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows, long numRows)
public static Throwable cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes, long dataSize)
public static Throwable notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe, scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables)
public static org.apache.spark.SparkUnsupportedOperationException executeCodePathUnsupportedError(String execName)
public static org.apache.spark.SparkUnsupportedOperationException cannotMergeClassWithOtherClassError(String className, String otherClass)
public static org.apache.spark.SparkUnsupportedOperationException continuousProcessingUnsupportedByDataSourceError(String sourceName)
public static Throwable failedToReadDataError(Throwable failureReason)
public static Throwable failedToGenerateEpochMarkerError(Throwable failureReason)
public static Throwable foreachWriterAbortedDueToTaskFailureError()
public static Throwable incorrectRampUpRate(long rowsPerSecond, long maxSeconds, long rampUpTimeSeconds)
public static Throwable incorrectEndOffset(long rowsPerSecond, long maxSeconds, long endSeconds)
public static Throwable failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, int keySize)
public static Throwable failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, String message)
public static org.apache.spark.SparkUnsupportedOperationException cannotPurgeAsBreakInternalStateError()
public static org.apache.spark.SparkUnsupportedOperationException cleanUpSourceFilesUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException latestOffsetNotCalledError()
public static Throwable legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath, String legacyCheckpointDir)
public static Throwable subprocessExitedError(int exitCode, org.apache.spark.util.CircularBuffer stderrBuffer, Throwable cause)
public static Throwable outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName, DataType dt)
public static org.apache.spark.SparkArrayIndexOutOfBoundsException invalidStartIndexError(int numRows, int startIndex)
public static org.apache.spark.SparkConcurrentModificationException concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className)
public static org.apache.spark.SparkUnsupportedOperationException doExecuteBroadcastNotImplementedError(String nodeName)
public static Throwable defaultDatabaseNotExistsError(String defaultDatabase)
public static Throwable databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB)
public static org.apache.spark.SparkSQLFeatureNotSupportedException commentOnTableUnsupportedError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedUpdateColumnNullabilityError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException renameColumnUnsupportedForOlderMySQLError()
public static SparkException failedToExecuteQueryError(Throwable e)
public static org.apache.spark.SparkUnsupportedOperationException nestedFieldUnsupportedError(String colName)
public static Throwable transformationsAndActionsNotInvokedByDriverError()
public static Throwable repeatedPivotsUnsupportedError(String clause, String operation)
public static Throwable pivotNotAfterGroupByUnsupportedError()
public static RuntimeException invalidAesKeyLengthError(int actualLength)
public static RuntimeException aesModeUnsupportedError(String mode, String padding)
public static RuntimeException aesCryptoError(String detailMessage)
public static RuntimeException invalidAesIvLengthError(String mode, int actualLength)
public static RuntimeException aesUnsupportedIv(String mode)
public static RuntimeException aesUnsupportedAad(String mode)
public static org.apache.spark.SparkUnsupportedOperationException hiveTableWithAnsiIntervalsError(org.apache.spark.sql.catalyst.TableIdentifier table)
public static Throwable cannotConvertOrcTimestampToTimestampNTZError()
public static Throwable cannotConvertOrcTimestampNTZToTimestampLTZError()
public static Throwable writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts, int maxDynamicPartitions, String maxDynamicPartitionsKey)
public static org.apache.spark.SparkIllegalArgumentException invalidNumberFormatError(DataType dataType, String input, String format)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedMultipleBucketTransformsError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedCommentNamespaceError(String namespace)
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedRemoveNamespaceCommentError(String namespace)
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedDropNamespaceError(String namespace)
public static org.apache.spark.SparkRuntimeException exceedMaxLimit(int limit)
public static ArithmeticException timestampAddOverflowError(long micros, int amount, String unit)
public static Throwable invalidBucketFile(String path)
public static Throwable multipleRowSubqueryError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static Throwable comparatorReturnsNull(String firstValue, String secondValue)
public static RuntimeException invalidPatternError(String funcName, String pattern, Throwable cause)
public static org.apache.spark.SparkIllegalArgumentException tooManyArrayElementsError(int numElements, int elementSize)
public static org.apache.spark.SparkIllegalArgumentException invalidEmptyLocationError(String location)
public static Throwable malformedProtobufMessageDetectedInMessageParsingError(Throwable e)
public static Throwable locationAlreadyExists(org.apache.spark.sql.catalyst.TableIdentifier tableId, org.apache.hadoop.fs.Path location)
public static Throwable cannotConvertCatalystValueToProtobufEnumTypeError(scala.collection.Seq<String> sqlColumn, String protobufColumn, String data, String enumString)
public static Throwable hllInvalidLgK(String function, int min, int max, String value)
public static Throwable hllInvalidInputSketchBuffer(String function)
public static Throwable hllUnionDifferentLgK(int left, int right, String function)
public static org.apache.spark.SparkRuntimeException mergeCardinalityViolationError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPurgePartitionError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPurgeTableError()
public static String toSQLId(String parts)
public static String toSQLId(scala.collection.Seq<String> parts)
public static String toSQLStmt(String text)
public static String toSQLConf(String conf)
public static String toSQLType(String text)
public static String toSQLType(org.apache.spark.sql.types.AbstractDataType t)
public static String toSQLValue(String value)
public static String toSQLValue(org.apache.spark.unsafe.types.UTF8String value)
public static String toSQLValue(short value)
public static String toSQLValue(int value)
public static String toSQLValue(long value)
public static String toSQLValue(float value)
public static String toSQLValue(double value)
public static String getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)
public static QueryContext[] getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)
public static String toSQLConfVal(String conf)
public static String toDSOption(String option)
public static String toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e)
public static String toSQLSchema(String schema)
public static String toSQLValue(Object v, DataType t)
public static org.apache.spark.SparkDateTimeException fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field, int actual, int expected, java.time.LocalDate candidate)
public static org.apache.spark.SparkUpgradeException failToParseDateTimeInNewParserError(String s, Throwable e)
public static org.apache.spark.SparkUpgradeException failToRecognizePatternAfterUpgradeError(String pattern, Throwable e)
public static org.apache.spark.SparkRuntimeException failToRecognizePatternError(String pattern, Throwable e)
public static org.apache.spark.SparkRuntimeException unreachableError(String err)
public static String unreachableError$default$1()
public static org.apache.spark.SparkDateTimeException invalidInputInCastToDatetimeError(org.apache.spark.unsafe.types.UTF8String value, DataType to, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkDateTimeException invalidInputInCastToDatetimeError(double value, DataType to, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static String arithmeticOverflowError$default$2()
public static org.apache.spark.sql.catalyst.trees.SQLQueryContext arithmeticOverflowError$default$3()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedArrowTypeError(org.apache.arrow.vector.types.pojo.ArrowType typeName)
public static org.apache.spark.SparkUnsupportedOperationException duplicatedFieldNameInArrowStructError(scala.collection.Seq<String> fieldNames)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedDataTypeError(DataType typeName)
public static Throwable userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt)
public static org.apache.spark.SparkUnsupportedOperationException cannotFindEncoderForTypeError(String typeName)
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInBeanClassError(Class<?> clazz)
public static org.apache.spark.SparkUnsupportedOperationException cannotFindConstructorForTypeError(String tpe)
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInClassError(String t)
public static org.apache.spark.SparkUnsupportedOperationException cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath)
public static org.apache.spark.SparkRuntimeException primaryConstructorNotFoundError(Class<?> cls)
public static org.apache.spark.SparkRuntimeException cannotGetOuterPointerForInnerClassError(Class<?> innerCls)