Package org.apache.spark.sql.errors
Class QueryExecutionErrors
Object
org.apache.spark.sql.errors.QueryExecutionErrors
Object for grouping error messages from (most) exceptions thrown during query execution.
This does not include exceptions thrown during the eager execution of commands, which are
grouped into
QueryCompilationErrors
.-
Constructor Summary
-
Method Summary
Modifier and TypeMethodDescriptionstatic org.apache.spark.SparkUnsupportedOperationException
addFilesWithAbsolutePathUnsupportedError
(String commitProtocol) static Throwable
static RuntimeException
aesCryptoError
(String detailMessage) static RuntimeException
aesModeUnsupportedError
(String mode, String padding) static RuntimeException
aesUnsupportedAad
(String mode) static RuntimeException
aesUnsupportedIv
(String mode) static org.apache.spark.SparkDateTimeException
static org.apache.spark.SparkDateTimeException
static org.apache.spark.SparkIllegalArgumentException
ansiIllegalArgumentError
(String message) static org.apache.spark.SparkArithmeticException
static String
static QueryContext
static org.apache.spark.SparkRuntimeException
arrayFunctionWithElementsExceedLimitError
(String prettyName, long numberOfElements) static org.apache.spark.SparkFileNotFoundException
batchMetadataFileNotFoundError
(org.apache.hadoop.fs.Path batchMetadataFile) static org.apache.spark.SparkArithmeticException
binaryArithmeticCauseOverflowError
(short eval1, String symbol, short eval2) static Throwable
bitPositionRangeError
(String funcName, int pos, int size) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotAcquireMemoryToBuildLongHashedRelationError
(long size, long got) static Throwable
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
static Throwable
cannotBroadcastTableOverMaxTableBytesError
(long maxBroadcastTableBytes, long dataSize) static Throwable
cannotBroadcastTableOverMaxTableRowsError
(long maxBroadcastTableRows, long numRows) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotCastError
(DataType from, DataType to) static Throwable
static ArithmeticException
cannotChangeDecimalPrecisionError
(Decimal value, int decimalPrecision, int decimalScale, QueryContext context) static Throwable
cannotClearOutputDirectoryError
(org.apache.hadoop.fs.Path staticPrefixPath) static Throwable
cannotClearPartitionDirectoryError
(org.apache.hadoop.fs.Path path) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkIllegalArgumentException
static Throwable
cannotConvertCatalystValueToProtobufEnumTypeError
(scala.collection.immutable.Seq<String> sqlColumn, String protobufColumn, String data, String enumString) static Throwable
static Throwable
static org.apache.spark.SparkRuntimeException
cannotCreateArrayWithElementsExceedLimitError
(long numElements, String additionalErrorMessage) static Throwable
static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkRuntimeException
cannotCreateParquetConverterForDataTypeError
(DataType t, String parquetType) static org.apache.spark.SparkRuntimeException
cannotCreateParquetConverterForDecimalTypeError
(DecimalType t, String parquetType) static org.apache.spark.SparkRuntimeException
cannotCreateParquetConverterForTypeError
(DecimalType t, String parquetType) static org.apache.spark.SparkRuntimeException
cannotCreateStagingDirError
(String message, IOException e) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotEvaluateExpressionError
(org.apache.spark.sql.catalyst.expressions.Expression expression) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotFetchTablesOfDatabaseError
(String dbName, Exception e) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
cannotFindEncoderForTypeError
(String typeName) static Throwable
cannotGenerateCodeForExpressionError
(org.apache.spark.sql.catalyst.expressions.Expression expression) static Throwable
cannotGenerateCodeForIncomparableTypeError
(String codeType, DataType dataType) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkRuntimeException
cannotGetOuterPointerForInnerClassError
(Class<?> innerCls) static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotInstantiateAbstractCatalogPluginClassError
(String name, String pluginClassName, Exception e) static Throwable
static Throwable
static org.apache.spark.SparkUnsupportedOperationException
cannotMergeClassWithOtherClassError
(String className, String otherClass) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
static org.apache.spark.SparkRuntimeException
cannotParseJsonArraysAsStructsError
(String recordStr) static org.apache.spark.SparkRuntimeException
cannotParseJSONFieldError
(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType) static org.apache.spark.SparkRuntimeException
cannotParseJSONFieldError
(String fieldName, String fieldValue, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkRuntimeException
cannotParseStringAsDataTypeError
(String recordStr, String fieldName, String fieldValue, DataType dataType) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
cannotReadCheckpoint
(String expectedVersion, String actualVersion) static Throwable
cannotReadFilesError
(Throwable e, String path) static Throwable
cannotReadFooterForFileError
(org.apache.hadoop.fs.Path file, Exception e) static Throwable
cannotRecognizeHiveTypeError
(org.apache.spark.sql.catalyst.parser.ParseException e, String fieldType, String fieldName) static org.apache.spark.SparkRuntimeException
cannotRemovePartitionDirError
(org.apache.hadoop.fs.Path partitionPath) static org.apache.spark.SparkUnsupportedOperationException
cannotRemoveReservedPropertyError
(String property) static Throwable
cannotRestorePermissionsForPathError
(org.apache.hadoop.fs.permission.FsPermission permission, org.apache.hadoop.fs.Path path) static org.apache.spark.SparkUnsupportedOperationException
cannotSafelyMergeSerdePropertiesError
(scala.collection.immutable.Map<String, String> props1, scala.collection.immutable.Map<String, String> props2, scala.collection.immutable.Set<String> conflictKeys) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkIllegalArgumentException
cannotSpecifyBothJdbcTableNameAndQueryError
(String jdbcTableName, String jdbcQueryString) static Throwable
cannotTerminateGeneratorError
(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkUnsupportedOperationException
cannotUseInvalidJavaIdentifierAsFieldNameError
(String fieldName, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) static org.apache.spark.SparkRuntimeException
static ArithmeticException
castingCauseOverflowError
(Object t, DataType from, DataType to) static ArithmeticException
castingCauseOverflowErrorInTableInsert
(DataType from, DataType to, String columnName) static Throwable
catalogFailToCallPublicNoArgConstructorError
(String name, String pluginClassName, Exception e) static Throwable
catalogFailToFindPublicNoArgConstructorError
(String name, String pluginClassName, Exception e) static Throwable
catalogNotFoundError
(String name) static Throwable
catalogPluginClassNotFoundForCatalogError
(String name, String pluginClassName, Exception e) static Throwable
catalogPluginClassNotImplementedError
(String name, String pluginClassName) static org.apache.spark.SparkRuntimeException
classHasUnexpectedSerializerError
(String clsName, org.apache.spark.sql.catalyst.expressions.Expression objSerializer) static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
codecNotAvailableError
(String codecName, String availableCodecs) static org.apache.spark.SparkSQLFeatureNotSupportedException
static Throwable
commitDeniedError
(int partId, long taskId, int attemptId, int stageId, int stageAttempt) static Throwable
comparatorReturnsNull
(String firstValue, String secondValue) static Throwable
compilerError
(org.codehaus.commons.compiler.CompileException e) static org.apache.spark.SparkConcurrentModificationException
static Throwable
static Throwable
concurrentStreamLogUpdate
(long batchId) static org.apache.spark.SparkRuntimeException
conflictingPartitionColumnNamesError
(scala.collection.immutable.Seq<String> distinctPartColLists, scala.collection.immutable.Seq<org.apache.hadoop.fs.Path> suspiciousPaths) static SparkException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
convertHiveTableToCatalogTableError
(SparkException e, String dbName, String tableName) static org.apache.spark.SparkRuntimeException
createArrayWithElementsExceedLimitError
(String prettyName, Object count) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
databaseNameConflictWithSystemPreservedDatabaseError
(String globalTempDB) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkClassNotFoundException
dataSourceNotFoundError
(String provider, Throwable error) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
decorrelateInnerQueryThroughPlanUnsupportedError
(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) static Throwable
defaultDatabaseNotExistsError
(String defaultDatabase) static ArithmeticException
divideByZeroError
(QueryContext context) static org.apache.spark.SparkUnsupportedOperationException
doExecuteBroadcastNotImplementedError
(String nodeName) static org.apache.spark.SparkUnsupportedOperationException
duplicatedFieldNameInArrowStructError
(scala.collection.immutable.Seq<String> fieldNames) static org.apache.spark.SparkRuntimeException
static Throwable
static Throwable
static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkRuntimeException
emittedRowsAreOlderThanWatermark
(long currentWatermark, long emittedRowEventTime) static org.apache.spark.SparkRuntimeException
emptyJsonFieldValueError
(DataType dataType) static org.apache.spark.SparkIllegalArgumentException
emptyOptionError
(String optionName) static Throwable
static Throwable
static org.apache.spark.SparkRuntimeException
exceedMapSizeLimitError
(int size) static org.apache.spark.SparkRuntimeException
exceedMaxLimit
(int limit) static Throwable
executeBroadcastTimeoutError
(long timeout, scala.Option<TimeoutException> ex) static org.apache.spark.SparkUnsupportedOperationException
executeCodePathUnsupportedError
(String execName) static org.apache.spark.SparkRuntimeException
expressionDecodingError
(Exception e, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) static org.apache.spark.SparkRuntimeException
expressionEncodingError
(Exception e, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) static Throwable
failedExecuteUserDefinedFunctionError
(String functionName, String inputTypes, String outputType, Throwable e) static Throwable
failedMergingSchemaError
(StructType leftSchema, StructType rightSchema, SparkException e) static Throwable
failedRenameTempFileError
(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath) static org.apache.spark.SparkRuntimeException
failedToCastValueToDataTypeForPartitionColumnError
(String value, DataType dataType, String columnName) static Throwable
failedToCommitStateFileError
(String providerClass, Throwable f) static String
static SparkException
static Throwable
failedToGenerateEpochMarkerError
(Throwable failureReason) static Throwable
failedToInstantiateConstructorForCatalogError
(String name, String pluginClassName, Exception e) static Throwable
failedToMergeIncompatibleSchemasError
(StructType left, StructType right, Throwable e) static Throwable
failedToPushRowIntoRowQueueError
(String rowQueue) static Throwable
failedToReadDataError
(Throwable failureReason) static Throwable
failedToReadDeltaFileKeySizeError
(org.apache.hadoop.fs.Path fileToRead, String clazz, int keySize) static Throwable
failedToReadDeltaFileNotExistsError
(org.apache.hadoop.fs.Path fileToRead, String clazz, Throwable f) static Throwable
failedToReadSnapshotFileKeySizeError
(org.apache.hadoop.fs.Path fileToRead, String clazz, int keySize) static Throwable
failedToReadSnapshotFileValueSizeError
(org.apache.hadoop.fs.Path fileToRead, String clazz, int valueSize) static Throwable
failedToReadStreamingStateFileError
(org.apache.hadoop.fs.Path fileToRead, Throwable f) static org.apache.spark.SparkRuntimeException
failToConvertValueToJsonError
(Object value, Class<?> cls, DataType dataType) static org.apache.spark.SparkUpgradeException
static org.apache.spark.SparkUpgradeException
failToRecognizePatternAfterUpgradeError
(String pattern, Throwable e) static org.apache.spark.SparkRuntimeException
failToRecognizePatternError
(String pattern, Throwable e) static org.apache.spark.SparkSecurityException
failToSetOriginalACLBackError
(String aclEntries, org.apache.hadoop.fs.Path path, Throwable e) static org.apache.spark.SparkIllegalArgumentException
fallbackV1RelationReportsInconsistentSchemaError
(StructType v2Schema, StructType v1Schema) static org.apache.spark.SparkRuntimeException
fieldCannotBeNullError
(int index, String fieldName) static String
fieldCannotBeNullMsg
(int index, String fieldName) static org.apache.spark.SparkDateTimeException
fieldDiffersFromDerivedLocalDateError
(ChronoField field, int actual, int expected, LocalDate candidate) static Throwable
fileLengthExceedsMaxLengthError
(org.apache.hadoop.fs.FileStatus status, int maxLength) static Throwable
fileNotExistError
(String path, Exception e) static Throwable
static org.apache.spark.SparkRuntimeException
foundDuplicateFieldInCaseInsensitiveModeError
(String requiredFieldName, String matchedOrcFields) static org.apache.spark.SparkRuntimeException
foundDuplicateFieldInFieldIdLookupModeError
(int requiredId, String matchedFields) static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkSQLFeatureNotSupportedException
getParentLoggerNotImplementedError
(String className) static org.apache.spark.SparkRuntimeException
static QueryContext[]
getQueryContext
(QueryContext context) static String
getSummary
(QueryContext sqlContext) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
hiveTableWithAnsiIntervalsError
(org.apache.spark.sql.catalyst.TableIdentifier table) static Throwable
hllInvalidInputSketchBuffer
(String function) static Throwable
hllInvalidLgK
(String function, int min, int max, int value) static Throwable
hllUnionDifferentLgK
(int left, int right, String function) static Throwable
illegalUrlError
(org.apache.spark.unsafe.types.UTF8String url, IllegalArgumentException e) static Throwable
static Throwable
incorrectEndOffset
(long rowsPerSecond, long maxSeconds, long endSeconds) static Throwable
incorrectRampUpRate
(long rowsPerSecond, long maxSeconds, long rampUpTimeSeconds) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
initialTypeNotTargetDataTypeError
(DataType dataType, String target) static org.apache.spark.SparkUnsupportedOperationException
initialTypeNotTargetDataTypesError
(DataType dataType) static org.apache.spark.SparkRuntimeException
static Throwable
static Throwable
internalCompilerError
(org.codehaus.commons.compiler.InternalCompilerException e) static ArithmeticException
intervalArithmeticOverflowError
(String message, String hint, QueryContext context) static ArithmeticException
intervalDividedByZeroError
(QueryContext context) static RuntimeException
invalidAesIvLengthError
(String mode, int actualLength) static RuntimeException
invalidAesKeyLengthError
(int actualLength) invalidArrayIndexError
(int index, int numElements, QueryContext context) invalidBitmapPositionError
(long bitPosition, long bitmapNumBytes) static Throwable
invalidBucketFile
(String path) static Throwable
static Throwable
invalidChangeLogReaderVersion
(long version) static Throwable
invalidChangeLogWriterVersion
(long version) static RuntimeException
invalidCharsetError
(String functionName, String charset) static Throwable
invalidDatetimeUnitError
(String functionName, String invalidValue) invalidElementAtIndexError
(int index, int numElements, QueryContext context) static Throwable
invalidExpressionEncoderError
(String encoderType) static org.apache.spark.SparkRuntimeException
invalidExternalTypeError
(String actualType, DataType expectedType, org.apache.spark.sql.catalyst.expressions.Expression childExpression) static DateTimeException
static RuntimeException
invalidIndexOfZeroError
(QueryContext context) static org.apache.spark.SparkDateTimeException
invalidInputInCastToDatetimeError
(double value, DataType to, QueryContext context) static org.apache.spark.SparkDateTimeException
invalidInputInCastToDatetimeError
(org.apache.spark.unsafe.types.UTF8String value, DataType to, QueryContext context) static org.apache.spark.SparkNumberFormatException
invalidInputInCastToNumberError
(DataType to, org.apache.spark.unsafe.types.UTF8String s, QueryContext context) static org.apache.spark.SparkIllegalArgumentException
invalidInputInConversionError
(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.unsafe.types.UTF8String fmt, String hint) static org.apache.spark.SparkRuntimeException
invalidInputSyntaxForBooleanError
(org.apache.spark.unsafe.types.UTF8String s, QueryContext context) static org.apache.spark.SparkIllegalArgumentException
invalidJdbcNumPartitionsError
(int n, String jdbcNumPartitions) static org.apache.spark.SparkIllegalArgumentException
invalidJdbcTxnIsolationLevelError
(String jdbcTxnIsolationLevel, String value) static Throwable
static org.apache.spark.SparkIllegalArgumentException
invalidLocationError
(String location, String errorClass, Throwable cause) static org.apache.spark.SparkIllegalArgumentException
invalidNamespaceNameError
(String[] namespace) static org.apache.spark.SparkIllegalArgumentException
invalidNumberFormatError
(DataType dataType, String input, String format) static org.apache.spark.SparkUnsupportedOperationException
static RuntimeException
invalidPatternError
(String funcName, String pattern, Throwable cause) static RuntimeException
invalidRegexGroupIndexError
(String funcName, int groupCount, int groupIndex) static org.apache.spark.SparkArrayIndexOutOfBoundsException
invalidStartIndexError
(int numRows, int startIndex) static org.apache.spark.SparkIllegalArgumentException
invalidUrlError
(org.apache.spark.unsafe.types.UTF8String url, URISyntaxException e) static org.apache.spark.SparkIllegalArgumentException
invalidUTF8StringError
(org.apache.spark.unsafe.types.UTF8String str) static Throwable
invalidVariantCast
(String value, DataType dataType) static Throwable
invalidVariantGetPath
(String path, String functionName) static Throwable
invalidWriterCommitMessageError
(String details) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
legacyCheckpointDirectoryExistsError
(org.apache.hadoop.fs.Path checkpointPath, String legacyCheckpointDir) static Throwable
legacyMetadataPathExistsError
(org.apache.hadoop.fs.Path metadataPath, org.apache.hadoop.fs.Path legacyMetadataPath) static RuntimeException
static org.apache.spark.SparkClassNotFoundException
loadHiveClientCausesNoClassDefFoundError
(NoClassDefFoundError cnf, scala.collection.immutable.Seq<URL> execJars, String key, InvocationTargetException e) static Throwable
locationAlreadyExists
(org.apache.spark.sql.catalyst.TableIdentifier tableId, org.apache.hadoop.fs.Path location) static RuntimeException
malformedCharacterCoding
(String functionName, String charset) static org.apache.spark.SparkRuntimeException
malformedCSVRecordError
(String badRecord) static Throwable
static Throwable
static Throwable
malformedRecordsDetectedInRecordParsingError
(String badRecord, Throwable e) static Throwable
static Throwable
static Throwable
static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkRuntimeException
mapSizeExceedArraySizeWhenZipMapError
(int size) static org.apache.spark.SparkRuntimeException
static Throwable
mergeUnsupportedByWindowFunctionError
(String funcName) static org.apache.spark.SparkRuntimeException
static Throwable
methodNotDeclaredError
(String name) static Throwable
methodNotFoundError
(Class<?> cls, String functionName, scala.collection.immutable.Seq<Class<?>> argClasses) static org.apache.spark.SparkUnsupportedOperationException
methodNotImplementedError
(String methodName) static org.apache.spark.SparkUnsupportedOperationException
microBatchUnsupportedByDataSourceError
(String srcName, String disabledSources, Table table) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkIllegalArgumentException
missingJdbcTableNameAndQueryError
(String jdbcTableName, String jdbcQueryString) static Throwable
multiActionAlterError
(String tableName) static Throwable
static org.apache.spark.SparkIllegalArgumentException
multiplePathsSpecifiedError
(scala.collection.immutable.Seq<String> allPaths) static Throwable
static org.apache.spark.SparkConcurrentModificationException
multiStreamingQueriesUsingPathConcurrentlyError
(String path, org.apache.hadoop.fs.FileAlreadyExistsException e) static org.apache.spark.SparkIllegalArgumentException
negativeValueUnexpectedError
(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkUnsupportedOperationException
nestedFieldUnsupportedError
(String colName) static SparkException
noDefaultForDataTypeError
(DataType dataType) static AnalysisException
nonTimeWindowNotSupportedInStreamingError
(scala.collection.immutable.Seq<String> windowFuncList, scala.collection.immutable.Seq<String> columnNameList, scala.collection.immutable.Seq<String> windowSpecList, org.apache.spark.sql.catalyst.trees.Origin origin) static Throwable
static Throwable
static Throwable
static Throwable
notEnoughMemoryToBuildAndBroadcastTableError
(OutOfMemoryError oe, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables) static Throwable
notEnoughMemoryToLoadStore
(String stateStoreId, String stateStoreProviderName, Throwable e) static org.apache.spark.SparkRuntimeException
notExpectedUnresolvedEncoderError
(org.apache.spark.sql.catalyst.expressions.AttributeReference attr) static org.apache.spark.SparkRuntimeException
notNullAssertViolation
(String walkedTypePath) static org.apache.spark.SparkRuntimeException
notOverrideExpectedMethodsError
(String className, String m1, String m2) static org.apache.spark.SparkUnsupportedOperationException
notPublicClassError
(String name) static org.apache.spark.SparkRuntimeException
static Throwable
notSupportTypeError
(DataType dataType) static org.apache.spark.SparkRuntimeException
static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>>
org.apache.spark.SparkRuntimeExceptiononceStrategyIdempotenceIsBrokenForBatchError
(String batchName, TreeType plan, TreeType reOptimized) static Throwable
onlySupportDataSourcesProvidingFileFormatError
(String providingClass) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkIllegalArgumentException
static String
ordinalNumber
(int i) static Throwable
outputDataTypeUnsupportedByNodeWithoutSerdeError
(String nodeName, DataType dt) static ArithmeticException
overflowInConvError
(QueryContext context) static ArithmeticException
static ArithmeticException
overflowInSumOfDecimalError
(QueryContext context) static Throwable
static org.apache.spark.SparkUnsupportedOperationException
pairUnsupportedAtFunctionError
(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1, org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2, String function) static org.apache.spark.SparkRuntimeException
paramExceedOneCharError
(String paramName) static Throwable
paramIsNotBooleanValueError
(String paramName) static org.apache.spark.SparkRuntimeException
paramIsNotIntegerError
(String paramName, String value) static Throwable
static Throwable
parquetColumnDataTypeMismatchError
(String path, String column, String expectedType, String actualType, Exception e) static org.apache.spark.SparkRuntimeException
partitionColumnNotFoundInSchemaError
(String col, StructType schema) static org.apache.spark.SparkIllegalArgumentException
partitionNumMismatchError
(int numFields, int schemaLen) static RuntimeException
pivotColumnUnsupportedError
(Object v, org.apache.spark.sql.catalyst.expressions.Expression expr) static Throwable
static org.apache.spark.SparkRuntimeException
primaryConstructorNotFoundError
(Class<?> cls) static org.apache.spark.SparkUnsupportedOperationException
static SparkException
pythonStreamingDataSourceRuntimeError
(String action, String message) static RuntimeException
raiseError
(org.apache.spark.unsafe.types.UTF8String errorClass, org.apache.spark.sql.catalyst.util.MapData errorParms) static Throwable
static org.apache.spark.SparkClassNotFoundException
removedClassInSpark2Error
(String className, Throwable e) static org.apache.spark.SparkFileAlreadyExistsException
renameAsExistsPathError
(org.apache.hadoop.fs.Path dstPath) static org.apache.spark.SparkSQLFeatureNotSupportedException
static Throwable
renamePathAsExistsPathError
(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath) static Throwable
renameSrcPathNotFoundError
(org.apache.hadoop.fs.Path srcPath) static Throwable
repeatedPivotsUnsupportedError
(String clause, String operation) static Throwable
requestedPartitionsMismatchTablePartitionsError
(org.apache.spark.sql.catalyst.catalog.CatalogTable table, scala.collection.immutable.Map<String, scala.Option<String>> partition) static org.apache.spark.SparkRuntimeException
resolveCannotHandleNestedSchema
(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
ruleIdNotFoundForRuleError
(String ruleName) static Throwable
saveModeUnsupportedError
(Object saveMode, boolean pathExists) static org.apache.spark.SparkClassNotFoundException
static org.apache.spark.SparkUpgradeException
sparkUpgradeInReadingDatesError
(String format, String config, String option) static org.apache.spark.SparkUpgradeException
sparkUpgradeInWritingDatesError
(String format, String config) static org.apache.spark.SparkNoSuchElementException
static org.apache.spark.SparkRuntimeException
statefulOperatorNotMatchInStateMetadataError
(scala.collection.immutable.Map<Object, String> opsInMetadataSeq, scala.collection.immutable.Map<Object, String> opsInCurBatchSeq) static Throwable
static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkUnsupportedOperationException
streamedOperatorUnsupportedByDataSourceError
(String className, String operator) static Throwable
subprocessExitedError
(int exitCode, org.apache.spark.util.CircularBuffer stderrBuffer, Throwable cause) static Throwable
taskFailedWhileWritingRowsError
(String path, Throwable cause) static ArithmeticException
timestampAddOverflowError
(long micros, int amount, String unit) static String
toDSOption
(String option) static org.apache.spark.SparkIllegalArgumentException
tooManyArrayElementsError
(long numElements, int maxRoundedArrayLength) static String
static String
toSQLConfVal
(String conf) static String
toSQLExpr
(org.apache.spark.sql.catalyst.expressions.Expression e) static String
static String
static String
static String
static String
toSQLType
(org.apache.spark.sql.types.AbstractDataType t) static String
toSQLValue
(double value) static String
toSQLValue
(float value) static String
toSQLValue
(int value) static String
toSQLValue
(long value) static String
toSQLValue
(short value) static String
toSQLValue
(Object v, DataType t) static String
toSQLValue
(String value) static String
toSQLValue
(org.apache.spark.unsafe.types.UTF8String value) static Throwable
static org.apache.spark.SparkUnsupportedOperationException
truncateMultiPartitionUnsupportedError
(String tableName) static Throwable
unableToCreateDatabaseAsFailedToCreateDirectoryError
(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, IOException e) static Throwable
unableToCreatePartitionPathError
(org.apache.hadoop.fs.Path partitionPath, IOException e) static Throwable
unableToCreateTableAsFailedToCreateDirectoryError
(String table, org.apache.hadoop.fs.Path defaultTableLocation, IOException e) static Throwable
unableToDeletePartitionPathError
(org.apache.hadoop.fs.Path partitionPath, IOException e) static Throwable
unableToDropDatabaseAsFailedToDeleteDirectoryError
(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, IOException e) static Throwable
unableToDropTableAsFailedToDeleteDirectoryError
(String table, org.apache.hadoop.fs.Path dir, IOException e) static Throwable
unableToRenamePartitionPathError
(org.apache.hadoop.fs.Path oldPartPath, IOException e) static Throwable
unableToRenameTableAsFailedToRenameDirectoryError
(String oldName, String newName, org.apache.hadoop.fs.Path oldDir, IOException e) static org.apache.spark.SparkIllegalArgumentException
static Throwable
unexpectedFileSize
(org.apache.hadoop.fs.Path dfsFile, File localFile, long expectedSize, long localFileSize) static org.apache.spark.SparkRuntimeException
unexpectedOperatorInCorrelatedSubquery
(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op, String pos) static Throwable
unexpectedStateStoreVersion
(long version) static org.apache.spark.SparkRuntimeException
unexpectedValueForLengthInFunctionError
(String prettyName, int length) static org.apache.spark.SparkRuntimeException
unexpectedValueForStartInFunctionError
(String prettyName) static org.apache.spark.SparkRuntimeException
static org.apache.spark.SparkIllegalArgumentException
unknownColumnError
(String unknownColumn) static org.apache.spark.SparkRuntimeException
unreachableError
(String err) static String
static org.apache.spark.SparkUnsupportedOperationException
unrecognizedCompressionSchemaTypeIDError
(int typeId) static Throwable
unrecognizedSqlTypeError
(String jdbcTypeId, String typeName) static Throwable
unreleasedThreadError
(String loggingId, String opType, String newAcquiredThreadInfo, String acquiredThreadInfo, long timeWaitedMs, String stackTraceOutput) static org.apache.spark.SparkIllegalArgumentException
static org.apache.spark.SparkUnsupportedOperationException
unsupportedArrowTypeError
(org.apache.arrow.vector.types.pojo.ArrowType typeName) static org.apache.spark.SparkSQLFeatureNotSupportedException
unsupportedCommentNamespaceError
(String namespace) static org.apache.spark.SparkUnsupportedOperationException
unsupportedDataTypeError
(DataType typeName) static org.apache.spark.SparkSQLFeatureNotSupportedException
unsupportedDropNamespaceError
(String namespace) static org.apache.spark.SparkRuntimeException
unsupportedFieldNameError
(String fieldName) static org.apache.spark.SparkUnsupportedOperationException
unsupportedHiveMetastoreVersionError
(String version, String key) static org.apache.spark.SparkSQLException
unsupportedJdbcTypeError
(String content) static org.apache.spark.SparkUnsupportedOperationException
static SparkException
unsupportedNaturalJoinTypeError
(org.apache.spark.sql.catalyst.plans.JoinType joinType) static Throwable
static org.apache.spark.SparkUnsupportedOperationException
unsupportedOperationForDataTypeError
(DataType dataType) static org.apache.spark.SparkUnsupportedOperationException
unsupportedOutputModeForStreamingOperationError
(OutputMode outputMode, String operation) static org.apache.spark.SparkUnsupportedOperationException
unsupportedPartitionTransformError
(Transform transform) static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkUnsupportedOperationException
static org.apache.spark.SparkSQLFeatureNotSupportedException
unsupportedRemoveNamespaceCommentError
(String namespace) static AnalysisException
unsupportedStreamingOperatorWithoutWatermark
(String outputMode, String statefulOperator) static Throwable
static org.apache.spark.SparkSQLFeatureNotSupportedException
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
static Throwable
static org.apache.spark.SparkUnsupportedOperationException
static Throwable
variantSizeLimitError
(int sizeLimit, String functionName) static Throwable
writePartitionExceedConfigSizeWhenDynamicPartitionError
(int numWrittenParts, int maxDynamicPartitions, String maxDynamicPartitionsKey) static org.apache.spark.SparkUnsupportedOperationException
static Throwable
writingJobFailedError
(Throwable cause)
-
Constructor Details
-
QueryExecutionErrors
public QueryExecutionErrors()
-
-
Method Details
-
cannotEvaluateExpressionError
public static Throwable cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) -
cannotGenerateCodeForExpressionError
public static Throwable cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) -
cannotTerminateGeneratorError
public static Throwable cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator) -
castingCauseOverflowError
-
castingCauseOverflowErrorInTableInsert
public static ArithmeticException castingCauseOverflowErrorInTableInsert(DataType from, DataType to, String columnName) -
cannotChangeDecimalPrecisionError
public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, QueryContext context) -
invalidInputSyntaxForBooleanError
public static org.apache.spark.SparkRuntimeException invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s, QueryContext context) -
invalidInputInCastToNumberError
public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, QueryContext context) -
invalidInputInConversionError
-
cannotCastFromNullTypeError
-
cannotCastError
-
cannotParseDecimalError
-
failedExecuteUserDefinedFunctionError
-
divideByZeroError
-
intervalDividedByZeroError
-
invalidUTF8StringError
public static org.apache.spark.SparkIllegalArgumentException invalidUTF8StringError(org.apache.spark.unsafe.types.UTF8String str) -
invalidArrayIndexError
public static ArrayIndexOutOfBoundsException invalidArrayIndexError(int index, int numElements, QueryContext context) -
invalidElementAtIndexError
public static ArrayIndexOutOfBoundsException invalidElementAtIndexError(int index, int numElements, QueryContext context) -
invalidBitmapPositionError
public static ArrayIndexOutOfBoundsException invalidBitmapPositionError(long bitPosition, long bitmapNumBytes) -
invalidFractionOfSecondError
-
ansiDateTimeParseError
-
ansiDateTimeError
-
ansiIllegalArgumentError
public static org.apache.spark.SparkIllegalArgumentException ansiIllegalArgumentError(String message) -
overflowInSumOfDecimalError
-
overflowInIntegralDivideError
-
overflowInConvError
-
mapSizeExceedArraySizeWhenZipMapError
public static org.apache.spark.SparkRuntimeException mapSizeExceedArraySizeWhenZipMapError(int size) -
literalTypeUnsupportedError
-
pivotColumnUnsupportedError
public static RuntimeException pivotColumnUnsupportedError(Object v, org.apache.spark.sql.catalyst.expressions.Expression expr) -
noDefaultForDataTypeError
-
orderedOperationUnsupportedByDataTypeError
public static org.apache.spark.SparkIllegalArgumentException orderedOperationUnsupportedByDataTypeError(DataType dataType) -
orderedOperationUnsupportedByDataTypeError
public static org.apache.spark.SparkIllegalArgumentException orderedOperationUnsupportedByDataTypeError(String dataType) -
invalidRegexGroupIndexError
public static RuntimeException invalidRegexGroupIndexError(String funcName, int groupCount, int groupIndex) -
invalidUrlError
public static org.apache.spark.SparkIllegalArgumentException invalidUrlError(org.apache.spark.unsafe.types.UTF8String url, URISyntaxException e) -
illegalUrlError
public static Throwable illegalUrlError(org.apache.spark.unsafe.types.UTF8String url, IllegalArgumentException e) -
mergeUnsupportedByWindowFunctionError
-
negativeValueUnexpectedError
public static org.apache.spark.SparkIllegalArgumentException negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression) -
addNewFunctionMismatchedWithFunctionError
-
cannotGenerateCodeForIncomparableTypeError
-
cannotInterpolateClassIntoCodeBlockError
-
customCollectionClsNotResolvedError
public static org.apache.spark.SparkUnsupportedOperationException customCollectionClsNotResolvedError() -
classUnsupportedByMapObjectsError
public static org.apache.spark.SparkRuntimeException classUnsupportedByMapObjectsError(Class<?> cls) -
nullAsMapKeyNotAllowedError
public static org.apache.spark.SparkRuntimeException nullAsMapKeyNotAllowedError() -
methodNotDeclaredError
-
methodNotFoundError
-
constructorNotFoundError
-
unsupportedNaturalJoinTypeError
public static SparkException unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType) -
notExpectedUnresolvedEncoderError
public static org.apache.spark.SparkRuntimeException notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr) -
invalidExpressionEncoderError
-
invalidExternalTypeError
-
notOverrideExpectedMethodsError
-
failToConvertValueToJsonError
-
resolveCannotHandleNestedSchema
public static org.apache.spark.SparkRuntimeException resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) -
inputExternalRowCannotBeNullError
public static org.apache.spark.SparkRuntimeException inputExternalRowCannotBeNullError() -
fieldCannotBeNullMsg
-
fieldCannotBeNullError
public static org.apache.spark.SparkRuntimeException fieldCannotBeNullError(int index, String fieldName) -
unableToCreateDatabaseAsFailedToCreateDirectoryError
public static Throwable unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, IOException e) -
unableToDropDatabaseAsFailedToDeleteDirectoryError
public static Throwable unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, IOException e) -
unableToCreateTableAsFailedToCreateDirectoryError
public static Throwable unableToCreateTableAsFailedToCreateDirectoryError(String table, org.apache.hadoop.fs.Path defaultTableLocation, IOException e) -
unableToDeletePartitionPathError
public static Throwable unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath, IOException e) -
unableToDropTableAsFailedToDeleteDirectoryError
public static Throwable unableToDropTableAsFailedToDeleteDirectoryError(String table, org.apache.hadoop.fs.Path dir, IOException e) -
unableToRenameTableAsFailedToRenameDirectoryError
public static Throwable unableToRenameTableAsFailedToRenameDirectoryError(String oldName, String newName, org.apache.hadoop.fs.Path oldDir, IOException e) -
unableToCreatePartitionPathError
public static Throwable unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath, IOException e) -
unableToRenamePartitionPathError
public static Throwable unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath, IOException e) -
methodNotImplementedError
public static org.apache.spark.SparkUnsupportedOperationException methodNotImplementedError(String methodName) -
arithmeticOverflowError
public static org.apache.spark.SparkArithmeticException arithmeticOverflowError(ArithmeticException e) -
binaryArithmeticCauseOverflowError
public static org.apache.spark.SparkArithmeticException binaryArithmeticCauseOverflowError(short eval1, String symbol, short eval2) -
intervalArithmeticOverflowError
public static ArithmeticException intervalArithmeticOverflowError(String message, String hint, QueryContext context) -
failedToCompileMsg
-
internalCompilerError
public static Throwable internalCompilerError(org.codehaus.commons.compiler.InternalCompilerException e) -
compilerError
-
unsupportedTableChangeError
-
notADatasourceRDDPartitionError
-
dataPathNotSpecifiedError
public static org.apache.spark.SparkIllegalArgumentException dataPathNotSpecifiedError() -
createStreamingSourceNotSpecifySchemaError
public static org.apache.spark.SparkIllegalArgumentException createStreamingSourceNotSpecifySchemaError() -
streamedOperatorUnsupportedByDataSourceError
-
nonTimeWindowNotSupportedInStreamingError
public static AnalysisException nonTimeWindowNotSupportedInStreamingError(scala.collection.immutable.Seq<String> windowFuncList, scala.collection.immutable.Seq<String> columnNameList, scala.collection.immutable.Seq<String> windowSpecList, org.apache.spark.sql.catalyst.trees.Origin origin) -
multiplePathsSpecifiedError
public static org.apache.spark.SparkIllegalArgumentException multiplePathsSpecifiedError(scala.collection.immutable.Seq<String> allPaths) -
dataSourceNotFoundError
-
removedClassInSpark2Error
-
incompatibleDataSourceRegisterError
-
sparkUpgradeInReadingDatesError
-
sparkUpgradeInWritingDatesError
-
buildReaderUnsupportedForFileFormatError
public static org.apache.spark.SparkUnsupportedOperationException buildReaderUnsupportedForFileFormatError(String format) -
taskFailedWhileWritingRowsError
-
saveModeUnsupportedError
-
cannotClearOutputDirectoryError
-
cannotClearPartitionDirectoryError
-
failedToCastValueToDataTypeForPartitionColumnError
-
endOfStreamError
-
fallbackV1RelationReportsInconsistentSchemaError
public static org.apache.spark.SparkIllegalArgumentException fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema, StructType v1Schema) -
noRecordsFromEmptyDataReaderError
-
fileNotExistError
-
parquetColumnDataTypeMismatchError
-
cannotReadFilesError
-
cannotCreateColumnarReaderError
-
invalidNamespaceNameError
public static org.apache.spark.SparkIllegalArgumentException invalidNamespaceNameError(String[] namespace) -
unsupportedPartitionTransformError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPartitionTransformError(Transform transform) -
missingDatabaseLocationError
public static org.apache.spark.SparkIllegalArgumentException missingDatabaseLocationError() -
cannotRemoveReservedPropertyError
public static org.apache.spark.SparkUnsupportedOperationException cannotRemoveReservedPropertyError(String property) -
writingJobFailedError
-
commitDeniedError
public static Throwable commitDeniedError(int partId, long taskId, int attemptId, int stageId, int stageAttempt) -
cannotCreateJDBCTableWithPartitionsError
public static org.apache.spark.SparkUnsupportedOperationException cannotCreateJDBCTableWithPartitionsError() -
unsupportedUserSpecifiedSchemaError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedUserSpecifiedSchemaError() -
writeUnsupportedForBinaryFileDataSourceError
public static org.apache.spark.SparkUnsupportedOperationException writeUnsupportedForBinaryFileDataSourceError() -
fileLengthExceedsMaxLengthError
public static Throwable fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status, int maxLength) -
unsupportedFieldNameError
-
cannotSpecifyBothJdbcTableNameAndQueryError
-
missingJdbcTableNameAndQueryError
-
emptyOptionError
-
invalidJdbcTxnIsolationLevelError
-
cannotGetJdbcTypeError
-
unrecognizedSqlTypeError
-
unsupportedJdbcTypeError
-
unsupportedArrayElementTypeBasedOnBinaryError
public static org.apache.spark.SparkIllegalArgumentException unsupportedArrayElementTypeBasedOnBinaryError(DataType dt) -
nestedArraysUnsupportedError
public static org.apache.spark.SparkIllegalArgumentException nestedArraysUnsupportedError() -
cannotTranslateNonNullValueForFieldError
public static org.apache.spark.SparkIllegalArgumentException cannotTranslateNonNullValueForFieldError(int pos) -
invalidJdbcNumPartitionsError
public static org.apache.spark.SparkIllegalArgumentException invalidJdbcNumPartitionsError(int n, String jdbcNumPartitions) -
multiActionAlterError
-
unsupportedOperationForDataTypeError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationForDataTypeError(DataType dataType) -
inputFilterNotFullyConvertibleError
-
foundDuplicateFieldInCaseInsensitiveModeError
-
foundDuplicateFieldInFieldIdLookupModeError
public static org.apache.spark.SparkRuntimeException foundDuplicateFieldInFieldIdLookupModeError(int requiredId, String matchedFields) -
failedToMergeIncompatibleSchemasError
public static Throwable failedToMergeIncompatibleSchemasError(StructType left, StructType right, Throwable e) -
ddlUnsupportedTemporarilyError
public static org.apache.spark.SparkUnsupportedOperationException ddlUnsupportedTemporarilyError(String ddl) -
executeBroadcastTimeoutError
public static Throwable executeBroadcastTimeoutError(long timeout, scala.Option<TimeoutException> ex) -
cannotCompareCostWithTargetCostError
public static org.apache.spark.SparkIllegalArgumentException cannotCompareCostWithTargetCostError(String cost) -
notSupportTypeError
-
notSupportNonPrimitiveTypeError
public static org.apache.spark.SparkRuntimeException notSupportNonPrimitiveTypeError() -
useDictionaryEncodingWhenDictionaryOverflowError
-
endOfIteratorError
-
cannotAllocateMemoryToGrowBytesToBytesMapError
-
cannotAcquireMemoryToBuildLongHashedRelationError
-
cannotAcquireMemoryToBuildUnsafeHashedRelationError
-
rowLargerThan256MUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException rowLargerThan256MUnsupportedError() -
cannotBuildHashedRelationWithUniqueKeysExceededError
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationWithUniqueKeysExceededError() -
cannotBuildHashedRelationLargerThan8GError
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationLargerThan8GError() -
failedToPushRowIntoRowQueueError
-
unexpectedWindowFunctionFrameError
public static org.apache.spark.SparkRuntimeException unexpectedWindowFunctionFrameError(String frame) -
cannotParseStatisticAsPercentileError
public static org.apache.spark.SparkIllegalArgumentException cannotParseStatisticAsPercentileError(String stats, NumberFormatException e) -
statisticNotRecognizedError
public static org.apache.spark.SparkIllegalArgumentException statisticNotRecognizedError(String stats) -
unknownColumnError
public static org.apache.spark.SparkIllegalArgumentException unknownColumnError(String unknownColumn) -
unexpectedAccumulableUpdateValueError
public static org.apache.spark.SparkIllegalArgumentException unexpectedAccumulableUpdateValueError(Object o) -
exceedMapSizeLimitError
public static org.apache.spark.SparkRuntimeException exceedMapSizeLimitError(int size) -
duplicateMapKeyFoundError
-
mapDataKeyArrayLengthDiffersFromValueArrayLengthError
public static org.apache.spark.SparkRuntimeException mapDataKeyArrayLengthDiffersFromValueArrayLengthError() -
registeringStreamingQueryListenerError
-
concurrentQueryInstanceError
-
concurrentStreamLogUpdate
-
cannotParseJsonArraysAsStructsError
public static org.apache.spark.SparkRuntimeException cannotParseJsonArraysAsStructsError(String recordStr) -
cannotParseStringAsDataTypeError
-
emptyJsonFieldValueError
-
cannotParseJSONFieldError
public static org.apache.spark.SparkRuntimeException cannotParseJSONFieldError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType) -
cannotParseJSONFieldError
-
rootConverterReturnNullError
public static org.apache.spark.SparkRuntimeException rootConverterReturnNullError() -
paramExceedOneCharError
-
paramIsNotIntegerError
-
paramIsNotBooleanValueError
-
foundNullValueForNotNullableFieldError
public static org.apache.spark.SparkRuntimeException foundNullValueForNotNullableFieldError(String name) -
malformedCSVRecordError
-
expressionDecodingError
public static org.apache.spark.SparkRuntimeException expressionDecodingError(Exception e, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) -
expressionEncodingError
public static org.apache.spark.SparkRuntimeException expressionEncodingError(Exception e, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) -
classHasUnexpectedSerializerError
public static org.apache.spark.SparkRuntimeException classHasUnexpectedSerializerError(String clsName, org.apache.spark.sql.catalyst.expressions.Expression objSerializer) -
unsupportedOperandTypeForSizeFunctionError
-
unexpectedValueForStartInFunctionError
public static org.apache.spark.SparkRuntimeException unexpectedValueForStartInFunctionError(String prettyName) -
unexpectedValueForLengthInFunctionError
public static org.apache.spark.SparkRuntimeException unexpectedValueForLengthInFunctionError(String prettyName, int length) -
invalidIndexOfZeroError
-
arrayFunctionWithElementsExceedLimitError
public static org.apache.spark.SparkRuntimeException arrayFunctionWithElementsExceedLimitError(String prettyName, long numberOfElements) -
createArrayWithElementsExceedLimitError
-
initialTypeNotTargetDataTypeError
-
initialTypeNotTargetDataTypesError
public static org.apache.spark.SparkUnsupportedOperationException initialTypeNotTargetDataTypesError(DataType dataType) -
malformedRecordsDetectedInSchemaInferenceError
-
malformedJSONError
-
malformedRecordsDetectedInSchemaInferenceError
-
decorrelateInnerQueryThroughPlanUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) -
methodCalledInAnalyzerNotAllowedError
public static org.apache.spark.SparkRuntimeException methodCalledInAnalyzerNotAllowedError() -
cannotSafelyMergeSerdePropertiesError
-
pairUnsupportedAtFunctionError
public static org.apache.spark.SparkUnsupportedOperationException pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1, org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2, String function) -
onceStrategyIdempotenceIsBrokenForBatchError
public static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> org.apache.spark.SparkRuntimeException onceStrategyIdempotenceIsBrokenForBatchError(String batchName, TreeType plan, TreeType reOptimized) -
ruleIdNotFoundForRuleError
-
cannotCreateArrayWithElementsExceedLimitError
public static org.apache.spark.SparkRuntimeException cannotCreateArrayWithElementsExceedLimitError(long numElements, String additionalErrorMessage) -
malformedRecordsDetectedInRecordParsingError
-
invalidKerberosConfigForHiveServer2Error
-
parentSparkUIToAttachTabNotFoundError
-
inferSchemaUnsupportedForHiveError
public static org.apache.spark.SparkUnsupportedOperationException inferSchemaUnsupportedForHiveError() -
requestedPartitionsMismatchTablePartitionsError
-
dynamicPartitionKeyNotAmongWrittenPartitionPathsError
-
cannotRemovePartitionDirError
public static org.apache.spark.SparkRuntimeException cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath) -
cannotCreateStagingDirError
public static org.apache.spark.SparkRuntimeException cannotCreateStagingDirError(String message, IOException e) -
serDeInterfaceNotFoundError
public static org.apache.spark.SparkClassNotFoundException serDeInterfaceNotFoundError(NoClassDefFoundError e) -
convertHiveTableToCatalogTableError
public static Throwable convertHiveTableToCatalogTableError(SparkException e, String dbName, String tableName) -
cannotRecognizeHiveTypeError
-
getTablesByTypeUnsupportedByHiveVersionError
public static org.apache.spark.SparkUnsupportedOperationException getTablesByTypeUnsupportedByHiveVersionError() -
invalidPartitionFilterError
public static org.apache.spark.SparkUnsupportedOperationException invalidPartitionFilterError() -
getPartitionMetadataByFilterError
-
unsupportedHiveMetastoreVersionError
-
loadHiveClientCausesNoClassDefFoundError
public static org.apache.spark.SparkClassNotFoundException loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf, scala.collection.immutable.Seq<URL> execJars, String key, InvocationTargetException e) -
cannotFetchTablesOfDatabaseError
-
renamePathAsExistsPathError
public static Throwable renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath) -
renameAsExistsPathError
public static org.apache.spark.SparkFileAlreadyExistsException renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath) -
renameSrcPathNotFoundError
-
failedRenameTempFileError
public static Throwable failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath) -
legacyMetadataPathExistsError
public static Throwable legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath, org.apache.hadoop.fs.Path legacyMetadataPath) -
partitionColumnNotFoundInSchemaError
public static org.apache.spark.SparkRuntimeException partitionColumnNotFoundInSchemaError(String col, StructType schema) -
stateNotDefinedOrAlreadyRemovedError
-
statefulOperatorNotMatchInStateMetadataError
-
cannotSetTimeoutDurationError
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutDurationError() -
cannotGetEventTimeWatermarkError
public static org.apache.spark.SparkUnsupportedOperationException cannotGetEventTimeWatermarkError() -
cannotSetTimeoutTimestampError
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutTimestampError() -
batchMetadataFileNotFoundError
public static org.apache.spark.SparkFileNotFoundException batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile) -
multiStreamingQueriesUsingPathConcurrentlyError
public static org.apache.spark.SparkConcurrentModificationException multiStreamingQueriesUsingPathConcurrentlyError(String path, org.apache.hadoop.fs.FileAlreadyExistsException e) -
addFilesWithAbsolutePathUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException addFilesWithAbsolutePathUnsupportedError(String commitProtocol) -
microBatchUnsupportedByDataSourceError
-
cannotExecuteStreamingRelationExecError
public static org.apache.spark.SparkUnsupportedOperationException cannotExecuteStreamingRelationExecError() -
unsupportedOutputModeForStreamingOperationError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOutputModeForStreamingOperationError(OutputMode outputMode, String operation) -
pythonStreamingDataSourceRuntimeError
-
invalidCatalogNameError
-
catalogNotFoundError
-
catalogPluginClassNotImplementedError
-
catalogPluginClassNotFoundForCatalogError
-
catalogFailToFindPublicNoArgConstructorError
-
catalogFailToCallPublicNoArgConstructorError
-
cannotInstantiateAbstractCatalogPluginClassError
-
failedToInstantiateConstructorForCatalogError
-
noSuchElementExceptionError
-
sqlConfigNotFoundError
-
cannotMutateReadOnlySQLConfError
public static org.apache.spark.SparkUnsupportedOperationException cannotMutateReadOnlySQLConfError() -
cannotCloneOrCopyReadOnlySQLConfError
public static org.apache.spark.SparkUnsupportedOperationException cannotCloneOrCopyReadOnlySQLConfError() -
cannotGetSQLConfInSchedulerEventLoopThreadError
public static org.apache.spark.SparkRuntimeException cannotGetSQLConfInSchedulerEventLoopThreadError() -
notPublicClassError
-
primitiveTypesNotSupportedError
public static org.apache.spark.SparkUnsupportedOperationException primitiveTypesNotSupportedError() -
onlySupportDataSourcesProvidingFileFormatError
-
cannotRestorePermissionsForPathError
public static Throwable cannotRestorePermissionsForPathError(org.apache.hadoop.fs.permission.FsPermission permission, org.apache.hadoop.fs.Path path) -
failToSetOriginalACLBackError
-
multiFailuresInStageMaterializationError
-
unrecognizedCompressionSchemaTypeIDError
public static org.apache.spark.SparkUnsupportedOperationException unrecognizedCompressionSchemaTypeIDError(int typeId) -
getParentLoggerNotImplementedError
public static org.apache.spark.SparkSQLFeatureNotSupportedException getParentLoggerNotImplementedError(String className) -
cannotCreateParquetConverterForTypeError
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForTypeError(DecimalType t, String parquetType) -
cannotCreateParquetConverterForDecimalTypeError
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForDecimalTypeError(DecimalType t, String parquetType) -
cannotCreateParquetConverterForDataTypeError
-
cannotAddMultiPartitionsOnNonatomicPartitionTableError
public static org.apache.spark.SparkUnsupportedOperationException cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName) -
userSpecifiedSchemaUnsupportedByDataSourceError
public static org.apache.spark.SparkUnsupportedOperationException userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider) -
cannotDropMultiPartitionsOnNonatomicPartitionTableError
public static org.apache.spark.SparkUnsupportedOperationException cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName) -
truncateMultiPartitionUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException truncateMultiPartitionUnsupportedError(String tableName) -
overwriteTableByUnsupportedExpressionError
-
dynamicPartitionOverwriteUnsupportedByTableError
-
failedMergingSchemaError
public static Throwable failedMergingSchemaError(StructType leftSchema, StructType rightSchema, SparkException e) -
cannotBroadcastTableOverMaxTableRowsError
public static Throwable cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows, long numRows) -
cannotBroadcastTableOverMaxTableBytesError
public static Throwable cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes, long dataSize) -
notEnoughMemoryToBuildAndBroadcastTableError
public static Throwable notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe, scala.collection.immutable.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables) -
executeCodePathUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException executeCodePathUnsupportedError(String execName) -
cannotMergeClassWithOtherClassError
-
continuousProcessingUnsupportedByDataSourceError
public static org.apache.spark.SparkUnsupportedOperationException continuousProcessingUnsupportedByDataSourceError(String sourceName) -
failedToReadDataError
-
failedToGenerateEpochMarkerError
-
foreachWriterAbortedDueToTaskFailureError
-
incorrectRampUpRate
public static Throwable incorrectRampUpRate(long rowsPerSecond, long maxSeconds, long rampUpTimeSeconds) -
incorrectEndOffset
-
failedToReadDeltaFileKeySizeError
-
failedToReadDeltaFileNotExistsError
-
failedToReadSnapshotFileKeySizeError
-
failedToReadSnapshotFileValueSizeError
-
failedToReadStreamingStateFileError
-
failedToCommitStateFileError
-
cannotPurgeAsBreakInternalStateError
public static org.apache.spark.SparkUnsupportedOperationException cannotPurgeAsBreakInternalStateError() -
cleanUpSourceFilesUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException cleanUpSourceFilesUnsupportedError() -
latestOffsetNotCalledError
public static org.apache.spark.SparkUnsupportedOperationException latestOffsetNotCalledError() -
legacyCheckpointDirectoryExistsError
-
subprocessExitedError
-
outputDataTypeUnsupportedByNodeWithoutSerdeError
-
invalidStartIndexError
public static org.apache.spark.SparkArrayIndexOutOfBoundsException invalidStartIndexError(int numRows, int startIndex) -
concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError
public static org.apache.spark.SparkConcurrentModificationException concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className) -
doExecuteBroadcastNotImplementedError
public static org.apache.spark.SparkUnsupportedOperationException doExecuteBroadcastNotImplementedError(String nodeName) -
defaultDatabaseNotExistsError
-
databaseNameConflictWithSystemPreservedDatabaseError
-
commentOnTableUnsupportedError
public static org.apache.spark.SparkSQLFeatureNotSupportedException commentOnTableUnsupportedError() -
unsupportedUpdateColumnNullabilityError
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedUpdateColumnNullabilityError() -
renameColumnUnsupportedForOlderMySQLError
public static org.apache.spark.SparkSQLFeatureNotSupportedException renameColumnUnsupportedForOlderMySQLError() -
failedToExecuteQueryError
-
nestedFieldUnsupportedError
public static org.apache.spark.SparkUnsupportedOperationException nestedFieldUnsupportedError(String colName) -
transformationsAndActionsNotInvokedByDriverError
-
repeatedPivotsUnsupportedError
-
pivotNotAfterGroupByUnsupportedError
-
invalidAesKeyLengthError
-
aesModeUnsupportedError
-
aesCryptoError
-
invalidAesIvLengthError
-
aesUnsupportedIv
-
aesUnsupportedAad
-
hiveTableWithAnsiIntervalsError
public static org.apache.spark.SparkUnsupportedOperationException hiveTableWithAnsiIntervalsError(org.apache.spark.sql.catalyst.TableIdentifier table) -
cannotConvertOrcTimestampToTimestampNTZError
-
cannotConvertOrcTimestampNTZToTimestampLTZError
-
writePartitionExceedConfigSizeWhenDynamicPartitionError
-
invalidNumberFormatError
-
unsupportedMultipleBucketTransformsError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedMultipleBucketTransformsError() -
unsupportedCommentNamespaceError
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedCommentNamespaceError(String namespace) -
unsupportedRemoveNamespaceCommentError
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedRemoveNamespaceCommentError(String namespace) -
unsupportedDropNamespaceError
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedDropNamespaceError(String namespace) -
exceedMaxLimit
public static org.apache.spark.SparkRuntimeException exceedMaxLimit(int limit) -
timestampAddOverflowError
-
invalidBucketFile
-
multipleRowScalarSubqueryError
-
comparatorReturnsNull
-
invalidPatternError
public static RuntimeException invalidPatternError(String funcName, String pattern, Throwable cause) -
tooManyArrayElementsError
public static org.apache.spark.SparkIllegalArgumentException tooManyArrayElementsError(long numElements, int maxRoundedArrayLength) -
invalidLocationError
-
malformedProtobufMessageDetectedInMessageParsingError
-
locationAlreadyExists
public static Throwable locationAlreadyExists(org.apache.spark.sql.catalyst.TableIdentifier tableId, org.apache.hadoop.fs.Path location) -
cannotConvertCatalystValueToProtobufEnumTypeError
-
unreleasedThreadError
-
cannotReadCheckpoint
-
unexpectedFileSize
-
unexpectedStateStoreVersion
-
invalidChangeLogReaderVersion
-
invalidChangeLogWriterVersion
-
notEnoughMemoryToLoadStore
-
cannotLoadStore
-
hllInvalidLgK
-
hllInvalidInputSketchBuffer
-
hllUnionDifferentLgK
-
mergeCardinalityViolationError
public static org.apache.spark.SparkRuntimeException mergeCardinalityViolationError() -
unsupportedPurgePartitionError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPurgePartitionError() -
unsupportedPurgeTableError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPurgeTableError() -
raiseError
public static RuntimeException raiseError(org.apache.spark.unsafe.types.UTF8String errorClass, org.apache.spark.sql.catalyst.util.MapData errorParms) -
bitPositionRangeError
-
variantSizeLimitError
-
invalidVariantCast
-
invalidVariantGetPath
-
malformedVariant
-
invalidCharsetError
-
malformedCharacterCoding
-
invalidWriterCommitMessageError
-
codecNotAvailableError
-
partitionNumMismatchError
public static org.apache.spark.SparkIllegalArgumentException partitionNumMismatchError(int numFields, int schemaLen) -
emittedRowsAreOlderThanWatermark
public static org.apache.spark.SparkRuntimeException emittedRowsAreOlderThanWatermark(long currentWatermark, long emittedRowEventTime) -
notNullAssertViolation
-
invalidDatetimeUnitError
-
unsupportedStreamingOperatorWithoutWatermark
public static AnalysisException unsupportedStreamingOperatorWithoutWatermark(String outputMode, String statefulOperator) -
conflictingPartitionColumnNamesError
public static org.apache.spark.SparkRuntimeException conflictingPartitionColumnNamesError(scala.collection.immutable.Seq<String> distinctPartColLists, scala.collection.immutable.Seq<org.apache.hadoop.fs.Path> suspiciousPaths) -
toSQLId
-
toSQLId
-
toSQLStmt
-
toSQLConf
-
toSQLType
-
toSQLType
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
getSummary
-
getQueryContext
-
toDSOption
-
toSQLConfVal
-
toSQLExpr
-
toSQLValue
-
ordinalNumber
-
fieldDiffersFromDerivedLocalDateError
public static org.apache.spark.SparkDateTimeException fieldDiffersFromDerivedLocalDateError(ChronoField field, int actual, int expected, LocalDate candidate) -
failToParseDateTimeInNewParserError
-
stateStoreHandleNotInitialized
public static org.apache.spark.SparkRuntimeException stateStoreHandleNotInitialized() -
failToRecognizePatternAfterUpgradeError
-
failToRecognizePatternError
-
unreachableError
-
unreachableError$default$1
-
invalidInputInCastToDatetimeError
public static org.apache.spark.SparkDateTimeException invalidInputInCastToDatetimeError(org.apache.spark.unsafe.types.UTF8String value, DataType to, QueryContext context) -
invalidInputInCastToDatetimeError
public static org.apache.spark.SparkDateTimeException invalidInputInCastToDatetimeError(double value, DataType to, QueryContext context) -
arithmeticOverflowError$default$2
-
arithmeticOverflowError$default$3
-
unsupportedArrowTypeError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedArrowTypeError(org.apache.arrow.vector.types.pojo.ArrowType typeName) -
duplicatedFieldNameInArrowStructError
public static org.apache.spark.SparkUnsupportedOperationException duplicatedFieldNameInArrowStructError(scala.collection.immutable.Seq<String> fieldNames) -
unsupportedDataTypeError
public static org.apache.spark.SparkUnsupportedOperationException unsupportedDataTypeError(DataType typeName) -
userDefinedTypeNotAnnotatedAndRegisteredError
-
cannotFindEncoderForTypeError
public static org.apache.spark.SparkUnsupportedOperationException cannotFindEncoderForTypeError(String typeName) -
cannotHaveCircularReferencesInBeanClassError
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInBeanClassError(Class<?> clazz) -
cannotFindConstructorForTypeError
public static org.apache.spark.SparkUnsupportedOperationException cannotFindConstructorForTypeError(String tpe) -
cannotHaveCircularReferencesInClassError
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInClassError(String t) -
cannotUseInvalidJavaIdentifierAsFieldNameError
public static org.apache.spark.SparkUnsupportedOperationException cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) -
-
cannotGetOuterPointerForInnerClassError
public static org.apache.spark.SparkRuntimeException cannotGetOuterPointerForInnerClassError(Class<?> innerCls) -
cannotUseKryoSerialization
public static org.apache.spark.SparkRuntimeException cannotUseKryoSerialization() -
elementsOfTupleExceedLimitError
public static org.apache.spark.SparkUnsupportedOperationException elementsOfTupleExceedLimitError()
-