public class QueryExecutionErrors
extends Object
QueryCompilationErrors
.Constructor and Description |
---|
QueryExecutionErrors() |
Modifier and Type | Method and Description |
---|---|
static org.apache.spark.SparkUnsupportedOperationException |
addFilesWithAbsolutePathUnsupportedError(String commitProtocol) |
static org.apache.spark.SparkIllegalArgumentException |
addNewFunctionMismatchedWithFunctionError(String funcName) |
static RuntimeException |
aesCryptoError(String detailMessage) |
static RuntimeException |
aesModeUnsupportedError(String mode,
String padding) |
static org.apache.spark.SparkUnsupportedOperationException |
alterTableWithDropPartitionAndPurgeUnsupportedError() |
static org.apache.spark.SparkDateTimeException |
ansiDateTimeError(Exception e) |
static org.apache.spark.SparkDateTimeException |
ansiDateTimeParseError(Exception e) |
static IllegalArgumentException |
ansiIllegalArgumentError(IllegalArgumentException e) |
static org.apache.spark.SparkIllegalArgumentException |
ansiIllegalArgumentError(String message) |
static org.apache.spark.SparkArithmeticException |
arithmeticOverflowError(ArithmeticException e) |
static ArithmeticException |
arithmeticOverflowError(String message,
String hint,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkUnsupportedOperationException |
attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema) |
static org.apache.spark.SparkFileNotFoundException |
batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile) |
static org.apache.spark.SparkArithmeticException |
binaryArithmeticCauseOverflowError(short eval1,
String symbol,
short eval2) |
static org.apache.spark.SparkUnsupportedOperationException |
buildReaderUnsupportedForFileFormatError(String format) |
static Throwable |
cannotAcquireMemoryToBuildLongHashedRelationError(long size,
long got) |
static Throwable |
cannotAcquireMemoryToBuildUnsafeHashedRelationError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotAllocateMemoryToGrowBytesToBytesMapError() |
static Throwable |
cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes,
long dataSize) |
static Throwable |
cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows,
long numRows) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotBuildHashedRelationLargerThan8GError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotBuildHashedRelationWithUniqueKeysExceededError() |
static Throwable |
cannotCastError(DataType from,
DataType to) |
static Throwable |
cannotCastFromNullTypeError(DataType to) |
static ArithmeticException |
cannotChangeDecimalPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath) |
static Throwable |
cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotCloneOrCopyReadOnlySQLConfError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotCompareCostWithTargetCostError(String cost) |
static Throwable |
cannotConvertOrcTimestampNTZToTimestampLTZError() |
static Throwable |
cannotConvertOrcTimestampToTimestampNTZError() |
static org.apache.spark.SparkRuntimeException |
cannotCreateArrayWithElementsExceedLimitError(long numElements,
String additionalErrorMessage) |
static Throwable |
cannotCreateColumnarReaderError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotCreateJDBCTableWithPartitionsError() |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForDataTypeError(DataType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForDecimalTypeError(DecimalType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateParquetConverterForTypeError(DecimalType t,
String parquetType) |
static org.apache.spark.SparkRuntimeException |
cannotCreateStagingDirError(String message,
java.io.IOException e) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName) |
static Throwable |
cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotExecuteStreamingRelationExecError() |
static Throwable |
cannotFetchTablesOfDatabaseError(String dbName,
Exception e) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotFindConstructorForTypeError(String tpe) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotFindEncoderForTypeError(String typeName,
String docroot) |
static Throwable |
cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression) |
static org.apache.spark.SparkIllegalArgumentException |
cannotGenerateCodeForIncomparableTypeError(String codeType,
DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotGetEventTimeWatermarkError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotGetJdbcTypeError(DataType dt) |
static org.apache.spark.SparkRuntimeException |
cannotGetOuterPointerForInnerClassError(Class<?> innerCls) |
static org.apache.spark.SparkRuntimeException |
cannotGetSQLConfInSchedulerEventLoopThreadError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotHaveCircularReferencesInBeanClassError(Class<?> clazz) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotHaveCircularReferencesInClassError(String t) |
static Throwable |
cannotInstantiateAbstractCatalogPluginClassError(String name,
String pluginClassName,
Exception e) |
static org.apache.spark.SparkIllegalArgumentException |
cannotInterpolateClassIntoCodeBlockError(Object arg) |
static Throwable |
cannotLoadUserDefinedTypeError(String name,
String userClass) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotMergeClassWithOtherClassError(String className,
String otherClass) |
static Throwable |
cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale,
int rightScale) |
static Throwable |
cannotMergeIncompatibleDataTypesError(DataType left,
DataType right) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotMutateReadOnlySQLConfError() |
static Throwable |
cannotParseDecimalError() |
static org.apache.spark.SparkRuntimeException |
cannotParseJsonArraysAsStructsError() |
static org.apache.spark.SparkRuntimeException |
cannotParseJSONFieldError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken jsonType,
DataType dataType) |
static org.apache.spark.SparkRuntimeException |
cannotParseJSONFieldError(String fieldName,
String fieldValue,
com.fasterxml.jackson.core.JsonToken jsonType,
DataType dataType) |
static org.apache.spark.SparkIllegalArgumentException |
cannotParseStatisticAsPercentileError(String stats,
NumberFormatException e) |
static org.apache.spark.SparkRuntimeException |
cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser,
com.fasterxml.jackson.core.JsonToken token,
DataType dataType) |
static org.apache.spark.SparkRuntimeException |
cannotParseStringAsDataTypeError(String pattern,
String value,
DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotPurgeAsBreakInternalStateError() |
static Throwable |
cannotReadFilesError(Throwable e,
String path) |
static Throwable |
cannotReadFooterForFileError(org.apache.hadoop.fs.Path file,
Exception e) |
static Throwable |
cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e,
String fieldType,
String fieldName) |
static org.apache.spark.SparkRuntimeException |
cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotRemoveReservedPropertyError(String property) |
static Throwable |
cannotRestorePermissionsForPathError(org.apache.hadoop.fs.permission.FsPermission permission,
org.apache.hadoop.fs.Path path) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1,
scala.collection.immutable.Map<String,String> props2,
scala.collection.immutable.Set<String> conflictKeys) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSetTimeoutDurationError() |
static org.apache.spark.SparkUnsupportedOperationException |
cannotSetTimeoutTimestampError() |
static org.apache.spark.SparkIllegalArgumentException |
cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator) |
static org.apache.spark.SparkIllegalArgumentException |
cannotTranslateNonNullValueForFieldError(int pos) |
static org.apache.spark.SparkUnsupportedOperationException |
cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName,
org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath) |
static ArithmeticException |
castingCauseOverflowError(Object t,
DataType from,
DataType to) |
static ArithmeticException |
castingCauseOverflowErrorInTableInsert(DataType from,
DataType to,
String columnName) |
static Throwable |
catalogFailToCallPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogFailToFindPublicNoArgConstructorError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotFoundError(String name) |
static Throwable |
catalogPluginClassNotFoundForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
catalogPluginClassNotImplementedError(String name,
String pluginClassName) |
static org.apache.spark.SparkRuntimeException |
classHasUnexpectedSerializerError(String clsName,
org.apache.spark.sql.catalyst.expressions.Expression objSerializer) |
static org.apache.spark.SparkRuntimeException |
classUnsupportedByMapObjectsError(Class<?> cls) |
static org.apache.spark.SparkUnsupportedOperationException |
cleanUpSourceFilesUnsupportedError() |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
commentOnTableUnsupportedError() |
static Throwable |
commitDeniedError(int partId,
long taskId,
int attemptId,
int stageId,
int stageAttempt) |
static Throwable |
comparatorReturnsNull(String firstValue,
String secondValue) |
static Throwable |
compilerError(org.codehaus.commons.compiler.CompileException e) |
static org.apache.spark.SparkRuntimeException |
concatArraysWithElementsExceedLimitError(long numberOfElements) |
static org.apache.spark.SparkConcurrentModificationException |
concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className) |
static Throwable |
concurrentQueryInstanceError() |
static org.apache.spark.SparkRuntimeException |
constructorNotFoundError(String cls) |
static org.apache.spark.SparkUnsupportedOperationException |
continuousProcessingUnsupportedByDataSourceError(String sourceName) |
static Throwable |
convertHiveTableToCatalogTableError(SparkException e,
String dbName,
String tableName) |
static org.apache.spark.SparkRuntimeException |
createArrayWithElementsExceedLimitError(Object count) |
static org.apache.spark.SparkIllegalArgumentException |
createStreamingSourceNotSpecifySchemaError() |
static org.apache.spark.SparkUnsupportedOperationException |
customCollectionClsNotResolvedError() |
static Throwable |
databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB) |
static org.apache.spark.SparkIllegalArgumentException |
dataPathNotSpecifiedError() |
static org.apache.spark.SparkClassNotFoundException |
dataSourceNotFoundError(String provider,
Throwable error) |
static org.apache.spark.SparkUnsupportedOperationException |
dataTypeUnexpectedError(DataType dataType) |
static Throwable |
dataTypeUnsupportedError(String dataType,
String failure) |
static org.apache.spark.SparkUnsupportedOperationException |
dataTypeUnsupportedYetError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
ddlUnsupportedTemporarilyError(String ddl) |
static org.apache.spark.SparkArithmeticException |
decimalPrecisionExceedsMaxPrecisionError(int precision,
int maxPrecision) |
static org.apache.spark.SparkUnsupportedOperationException |
decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static Throwable |
defaultDatabaseNotExistsError(String defaultDatabase) |
static ArithmeticException |
divideByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkUnsupportedOperationException |
doExecuteBroadcastNotImplementedError(String nodeName) |
static org.apache.spark.SparkUnsupportedOperationException |
dropTableWithPurgeUnsupportedError() |
static org.apache.spark.SparkRuntimeException |
duplicateMapKeyFoundError(Object key) |
static Throwable |
dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key) |
static Throwable |
dynamicPartitionOverwriteUnsupportedByTableError(Table table) |
static org.apache.spark.SparkUnsupportedOperationException |
elementsOfTupleExceedLimitError() |
static org.apache.spark.SparkRuntimeException |
emptyJsonFieldValueError(DataType dataType) |
static org.apache.spark.SparkIllegalArgumentException |
emptyOptionError(String optionName) |
static Throwable |
endOfIteratorError() |
static Throwable |
endOfStreamError() |
static org.apache.spark.SparkRuntimeException |
exceedMapSizeLimitError(int size) |
static Throwable |
executeBroadcastTimeoutError(long timeout,
scala.Option<java.util.concurrent.TimeoutException> ex) |
static org.apache.spark.SparkUnsupportedOperationException |
executeCodePathUnsupportedError(String execName) |
static org.apache.spark.SparkRuntimeException |
expressionDecodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static org.apache.spark.SparkRuntimeException |
expressionEncodingError(Exception e,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions) |
static Throwable |
failedExecuteUserDefinedFunctionError(String funcCls,
String inputTypes,
String outputType,
Throwable e) |
static Throwable |
failedMergingSchemaError(StructType schema,
SparkException e) |
static org.apache.spark.SparkRuntimeException |
failedParsingStructTypeError(String raw) |
static Throwable |
failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static org.apache.spark.SparkRuntimeException |
failedToCastValueToDataTypeForPartitionColumnError(String value,
DataType dataType,
String columnName) |
static String |
failedToCompileMsg(Exception e) |
static SparkException |
failedToExecuteQueryError(Throwable e) |
static Throwable |
failedToGenerateEpochMarkerError(Throwable failureReason) |
static Throwable |
failedToInstantiateConstructorForCatalogError(String name,
String pluginClassName,
Exception e) |
static Throwable |
failedToMergeIncompatibleSchemasError(StructType left,
StructType right,
Throwable e) |
static Throwable |
failedToPushRowIntoRowQueueError(String rowQueue) |
static Throwable |
failedToReadDataError(Throwable failureReason) |
static Throwable |
failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
int keySize) |
static Throwable |
failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead,
String clazz,
String message) |
static org.apache.spark.SparkRuntimeException |
failToConvertValueToJsonError(Object value,
Class<?> cls,
DataType dataType) |
static Throwable |
failToParseDateTimeInNewParserError(String s,
Throwable e) |
static Throwable |
failToRecognizePatternAfterUpgradeError(String pattern,
Throwable e,
String docroot) |
static org.apache.spark.SparkRuntimeException |
failToRecognizePatternError(String pattern,
Throwable e,
String docroot) |
static org.apache.spark.SparkSecurityException |
failToSetOriginalACLBackError(String aclEntries,
org.apache.hadoop.fs.Path path,
Throwable e) |
static org.apache.spark.SparkIllegalArgumentException |
fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema,
StructType v1Schema) |
static org.apache.spark.SparkRuntimeException |
fieldCannotBeNullError(int index,
String fieldName) |
static String |
fieldCannotBeNullMsg(int index,
String fieldName) |
static org.apache.spark.SparkDateTimeException |
fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field,
int actual,
int expected,
java.time.LocalDate candidate) |
static org.apache.spark.SparkUnsupportedOperationException |
fieldIndexOnRowWithoutSchemaError() |
static Throwable |
fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status,
int maxLength) |
static org.apache.spark.SparkFileNotFoundException |
fileNotFoundError(java.io.FileNotFoundException e) |
static org.apache.spark.SparkRuntimeException |
flattenArraysWithElementsExceedLimitError(long numberOfElements) |
static Throwable |
foreachWriterAbortedDueToTaskFailureError() |
static org.apache.spark.SparkRuntimeException |
foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName,
String matchedOrcFields) |
static org.apache.spark.SparkRuntimeException |
foundDuplicateFieldInFieldIdLookupModeError(int requiredId,
String matchedFields) |
static org.apache.spark.SparkRuntimeException |
foundNullValueForNotNullableFieldError(String name) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
getParentLoggerNotImplementedError(String className) |
static org.apache.spark.SparkRuntimeException |
getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e) |
static QueryContext[] |
getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static String |
getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static org.apache.spark.SparkUnsupportedOperationException |
getTablesByTypeUnsupportedByHiveVersionError() |
static org.apache.spark.SparkUnsupportedOperationException |
hiveTableWithAnsiIntervalsError(String tableName) |
static Throwable |
illegalLocationClauseForViewPartitionError() |
static Throwable |
illegalUrlError(org.apache.spark.unsafe.types.UTF8String url) |
static Throwable |
incompatibleDataSourceRegisterError(Throwable e) |
static Throwable |
incorrectEndOffset(long rowsPerSecond,
long maxSeconds,
long endSeconds) |
static Throwable |
incorrectRampUpRate(long rowsPerSecond,
long maxSeconds,
long rampUpTimeSeconds) |
static org.apache.spark.SparkUnsupportedOperationException |
inferSchemaUnsupportedForHiveError() |
static org.apache.spark.SparkUnsupportedOperationException |
initialTypeNotTargetDataTypeError(DataType dataType,
String target) |
static org.apache.spark.SparkUnsupportedOperationException |
initialTypeNotTargetDataTypesError(DataType dataType) |
static org.apache.spark.SparkRuntimeException |
inputExternalRowCannotBeNullError() |
static Throwable |
inputFilterNotFullyConvertibleError(String owner) |
static Throwable |
internalCompilerError(org.codehaus.commons.compiler.InternalCompilerException e) |
static ArithmeticException |
intervalArithmeticOverflowError(String message,
String hint,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
intervalDividedByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static RuntimeException |
invalidAesKeyLengthError(int actualLength) |
static ArrayIndexOutOfBoundsException |
invalidArrayIndexError(int index,
int numElements,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
invalidBucketFile(String path) |
static Throwable |
invalidCatalogNameError(String name) |
static ArrayIndexOutOfBoundsException |
invalidElementAtIndexError(int index,
int numElements,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidEmptyLocationError(String location) |
static java.time.DateTimeException |
invalidFractionOfSecondError() |
static RuntimeException |
invalidIndexOfZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
invalidInputInCastToDatetimeError(Object value,
DataType from,
DataType to,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkNumberFormatException |
invalidInputInCastToNumberError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidInputInConversionError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.unsafe.types.UTF8String fmt,
String hint) |
static org.apache.spark.SparkRuntimeException |
invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
invalidJdbcNumPartitionsError(int n,
String jdbcNumPartitions) |
static org.apache.spark.SparkIllegalArgumentException |
invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel,
String value) |
static Throwable |
invalidKerberosConfigForHiveServer2Error() |
static org.apache.spark.SparkIllegalArgumentException |
invalidNamespaceNameError(String[] namespace) |
static org.apache.spark.SparkIllegalArgumentException |
invalidNumberFormatError(String valueType,
String input,
String format) |
static org.apache.spark.SparkUnsupportedOperationException |
invalidPartitionFilterError() |
static RuntimeException |
invalidPatternError(String funcName,
String pattern,
Throwable cause) |
static org.apache.spark.SparkArrayIndexOutOfBoundsException |
invalidStartIndexError(int numRows,
int startIndex) |
static org.apache.spark.SparkUnsupportedOperationException |
invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode) |
static org.apache.spark.SparkIllegalArgumentException |
invalidUrlError(org.apache.spark.unsafe.types.UTF8String url,
java.net.URISyntaxException e) |
static org.apache.spark.SparkUnsupportedOperationException |
latestOffsetNotCalledError() |
static Throwable |
legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath,
String legacyCheckpointDir) |
static Throwable |
legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath,
org.apache.hadoop.fs.Path legacyMetadataPath) |
static RuntimeException |
literalTypeUnsupportedError(Object v) |
static org.apache.spark.SparkClassNotFoundException |
loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf,
scala.collection.Seq<java.net.URL> execJars,
String key,
java.lang.reflect.InvocationTargetException e) |
static Throwable |
locationAlreadyExists(org.apache.spark.sql.catalyst.TableIdentifier tableId,
org.apache.hadoop.fs.Path location) |
static org.apache.spark.SparkRuntimeException |
malformedCSVRecordError(String badRecord) |
static Throwable |
malformedJSONError() |
static Throwable |
malformedProtobufMessageDetectedInMessageParsingError(Throwable e) |
static Throwable |
malformedRecordsDetectedInRecordParsingError(String badRecord,
org.apache.spark.sql.catalyst.util.BadRecordException e) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(DataType dataType) |
static Throwable |
malformedRecordsDetectedInSchemaInferenceError(Throwable e) |
static org.apache.spark.SparkRuntimeException |
mapDataKeyArrayLengthDiffersFromValueArrayLengthError() |
static org.apache.spark.SparkRuntimeException |
mapSizeExceedArraySizeWhenZipMapError(int size) |
static org.apache.spark.SparkUnsupportedOperationException |
mergeUnsupportedByWindowFunctionError() |
static org.apache.spark.SparkRuntimeException |
methodCalledInAnalyzerNotAllowedError() |
static Throwable |
methodNotDeclaredError(String name) |
static org.apache.spark.SparkUnsupportedOperationException |
methodNotImplementedError(String methodName) |
static org.apache.spark.SparkUnsupportedOperationException |
microBatchUnsupportedByDataSourceError(String srcName,
String disabledSources,
Table table) |
static org.apache.spark.SparkIllegalArgumentException |
missingDatabaseLocationError() |
static org.apache.spark.SparkIllegalArgumentException |
missingJdbcTableNameAndQueryError(String jdbcTableName,
String jdbcQueryString) |
static Throwable |
multiActionAlterError(String tableName) |
static Throwable |
multiFailuresInStageMaterializationError(Throwable error) |
static org.apache.spark.SparkIllegalArgumentException |
multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths) |
static Throwable |
multipleRowSubqueryError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkConcurrentModificationException |
multiStreamingQueriesUsingPathConcurrentlyError(String path,
org.apache.hadoop.fs.FileAlreadyExistsException e) |
static org.apache.spark.SparkIllegalArgumentException |
negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression) |
static org.apache.spark.SparkIllegalArgumentException |
nestedArraysUnsupportedError() |
static org.apache.spark.SparkUnsupportedOperationException |
nestedFieldUnsupportedError(String colName) |
static org.apache.spark.SparkRuntimeException |
noDefaultForDataTypeError(DataType dataType) |
static Throwable |
noRecordsFromEmptyDataReaderError() |
static Throwable |
noSuchElementExceptionError() |
static Throwable |
noSuchElementExceptionError(String key) |
static Throwable |
notADatasourceRDDPartitionError(Partition split) |
static Throwable |
notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe,
scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables) |
static org.apache.spark.SparkRuntimeException |
notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr) |
static org.apache.spark.SparkRuntimeException |
notOverrideExpectedMethodsError(String className,
String m1,
String m2) |
static org.apache.spark.SparkUnsupportedOperationException |
notPublicClassError(String name) |
static org.apache.spark.SparkRuntimeException |
notSupportNonPrimitiveTypeError() |
static Throwable |
notSupportTypeError(DataType dataType) |
static Throwable |
notUserDefinedTypeError(String name,
String userClass) |
static org.apache.spark.SparkRuntimeException |
nullAsMapKeyNotAllowedError() |
static org.apache.spark.SparkUnsupportedOperationException |
nullLiteralsCannotBeCastedError(String name) |
static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> |
onceStrategyIdempotenceIsBrokenForBatchError(String batchName,
TreeType plan,
TreeType reOptimized) |
static Throwable |
onlySupportDataSourcesProvidingFileFormatError(String providingClass) |
static org.apache.spark.SparkIllegalArgumentException |
orderedOperationUnsupportedByDataTypeError(DataType dataType) |
static org.apache.spark.SparkArithmeticException |
outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str) |
static Throwable |
outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName,
DataType dt) |
static ArithmeticException |
overflowInConvError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
overflowInIntegralDivideError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static ArithmeticException |
overflowInSumOfDecimalError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
overwriteTableByUnsupportedExpressionError(Table table) |
static org.apache.spark.SparkUnsupportedOperationException |
pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1,
org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2,
String function) |
static org.apache.spark.SparkRuntimeException |
paramExceedOneCharError(String paramName) |
static Throwable |
paramIsNotBooleanValueError(String paramName) |
static org.apache.spark.SparkRuntimeException |
paramIsNotIntegerError(String paramName,
String value) |
static Throwable |
parentSparkUIToAttachTabNotFoundError() |
static org.apache.spark.SparkRuntimeException |
partitionColumnNotFoundInSchemaError(String col,
StructType schema) |
static RuntimeException |
pivotColumnUnsupportedError(Object v,
DataType dataType) |
static Throwable |
pivotNotAfterGroupByUnsupportedError() |
static org.apache.spark.SparkRuntimeException |
primaryConstructorNotFoundError(Class<?> cls) |
static org.apache.spark.SparkUnsupportedOperationException |
primitiveTypesNotSupportedError() |
static org.apache.spark.SparkFileNotFoundException |
readCurrentFileNotFoundError(java.io.FileNotFoundException e) |
static org.apache.spark.SparkIllegalArgumentException |
regexGroupIndexExceedGroupCountError(int groupCount,
int groupIndex) |
static org.apache.spark.SparkIllegalArgumentException |
regexGroupIndexLessThanZeroError() |
static Throwable |
registeringStreamingQueryListenerError(Exception e) |
static org.apache.spark.SparkRuntimeException |
remoteOperationsUnsupportedError() |
static org.apache.spark.SparkClassNotFoundException |
removedClassInSpark2Error(String className,
Throwable e) |
static org.apache.spark.SparkFileAlreadyExistsException |
renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
renameColumnUnsupportedForOlderMySQLError() |
static Throwable |
renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath,
org.apache.hadoop.fs.Path dstPath) |
static Throwable |
renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath) |
static Throwable |
repeatedPivotsUnsupportedError(String clause,
String operation) |
static Throwable |
requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table,
scala.collection.immutable.Map<String,scala.Option<String>> partition) |
static org.apache.spark.SparkRuntimeException |
resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) |
static org.apache.spark.SparkRuntimeException |
rootConverterReturnNullError() |
static org.apache.spark.SparkUnsupportedOperationException |
rowLargerThan256MUnsupportedError() |
static Throwable |
ruleIdNotFoundForRuleError(String ruleName) |
static Throwable |
saveModeUnsupportedError(Object saveMode,
boolean pathExists) |
static org.apache.spark.SparkClassNotFoundException |
serDeInterfaceNotFoundError(NoClassDefFoundError e) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInReadingDatesError(String format,
String config,
String option) |
static org.apache.spark.SparkUpgradeException |
sparkUpgradeInWritingDatesError(String format,
String config) |
static Throwable |
stateNotDefinedOrAlreadyRemovedError() |
static org.apache.spark.SparkIllegalArgumentException |
statisticNotRecognizedError(String stats) |
static org.apache.spark.SparkUnsupportedOperationException |
streamedOperatorUnsupportedByDataSourceError(String className,
String operator) |
static Throwable |
subprocessExitedError(int exitCode,
org.apache.spark.util.CircularBuffer stderrBuffer,
Throwable cause) |
static Throwable |
taskFailedWhileWritingRowsError(String path,
Throwable cause) |
static ArithmeticException |
timestampAddOverflowError(long micros,
int amount,
String unit) |
static String |
toDSOption(String option) |
static org.apache.spark.SparkIllegalArgumentException |
tooManyArrayElementsError(int numElements,
int elementSize) |
static String |
toSQLConf(String conf) |
static String |
toSQLConfVal(String conf) |
static String |
toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e) |
static String |
toSQLId(scala.collection.Seq<String> parts) |
static String |
toSQLId(String parts) |
static String |
toSQLSchema(String schema) |
static String |
toSQLStmt(String text) |
static String |
toSQLType(org.apache.spark.sql.types.AbstractDataType t) |
static String |
toSQLValue(Object v,
DataType t) |
static Throwable |
transformationsAndActionsNotInvokedByDriverError() |
static org.apache.spark.SparkUnsupportedOperationException |
truncateMultiPartitionUnsupportedError(String tableName) |
static org.apache.spark.SparkIllegalArgumentException |
typeUnsupportedError(DataType dataType) |
static Throwable |
unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToCreateTableAsFailedToCreateDirectoryError(String table,
org.apache.hadoop.fs.Path defaultTableLocation,
java.io.IOException e) |
static Throwable |
unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath,
java.io.IOException e) |
static Throwable |
unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition,
java.io.IOException e) |
static Throwable |
unableToDropTableAsFailedToDeleteDirectoryError(String table,
org.apache.hadoop.fs.Path dir,
java.io.IOException e) |
static Throwable |
unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath,
java.io.IOException e) |
static Throwable |
unableToRenameTableAsFailedToRenameDirectoryError(String oldName,
String newName,
org.apache.hadoop.fs.Path oldDir,
java.io.IOException e) |
static org.apache.spark.SparkArithmeticException |
unaryMinusCauseOverflowError(int originValue) |
static org.apache.spark.SparkIllegalArgumentException |
unexpectedAccumulableUpdateValueError(Object o) |
static org.apache.spark.SparkRuntimeException |
unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op,
String pos) |
static org.apache.spark.SparkRuntimeException |
unexpectedValueForLengthInFunctionError(String prettyName) |
static org.apache.spark.SparkRuntimeException |
unexpectedValueForStartInFunctionError(String prettyName) |
static org.apache.spark.SparkRuntimeException |
unexpectedWindowFunctionFrameError(String frame) |
static org.apache.spark.SparkRuntimeException |
unionArrayWithElementsExceedLimitError(int length) |
static org.apache.spark.SparkIllegalArgumentException |
unknownColumnError(String unknownColumn) |
static org.apache.spark.SparkRuntimeException |
unreachableError(String err) |
static org.apache.spark.SparkUnsupportedOperationException |
unrecognizedCompressionSchemaTypeIDError(int typeId) |
static Throwable |
unrecognizedSqlTypeError(int sqlType) |
static ArithmeticException |
unscaledValueTooLargeForPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkIllegalArgumentException |
unsupportedArrayElementTypeBasedOnBinaryError(DataType dt) |
static org.apache.spark.SparkRuntimeException |
unsupportedArrayTypeError(Class<?> clazz) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedArrowTypeError(org.apache.arrow.vector.types.pojo.ArrowType typeName) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedCreateNamespaceCommentError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedDataTypeError(DataType typeName) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedDropNamespaceRestrictError() |
static org.apache.spark.SparkRuntimeException |
unsupportedEncoderError() |
static org.apache.spark.SparkRuntimeException |
unsupportedFieldNameError(String fieldName) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedHiveMetastoreVersionError(String version,
String key) |
static org.apache.spark.SparkRuntimeException |
unsupportedJavaTypeError(Class<?> clazz) |
static org.apache.spark.SparkSQLException |
unsupportedJdbcTypeError(String content) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedMultipleBucketTransformsError() |
static org.apache.spark.SparkRuntimeException |
unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperandTypeForSizeFunctionError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperationExceptionError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperationForDataTypeError(DataType dataType) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedPartitionTransformError(Transform transform) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedRemoveNamespaceCommentError() |
static org.apache.spark.SparkRuntimeException |
unsupportedRoundingMode(scala.Enumeration.Value roundMode) |
static Throwable |
unsupportedSchemaColumnConvertError(String filePath,
String column,
String logicalType,
String physicalType,
Exception e) |
static Throwable |
unsupportedTableChangeError(IllegalArgumentException e) |
static Throwable |
unsupportedTableWritesError(Identifier ident) |
static Throwable |
unsupportedTypeError(DataType dataType) |
static org.apache.spark.SparkSQLFeatureNotSupportedException |
unsupportedUpdateColumnNullabilityError() |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedUserSpecifiedSchemaError() |
static Throwable |
useDictionaryEncodingWhenDictionaryOverflowError() |
static Throwable |
userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt) |
static org.apache.spark.SparkUnsupportedOperationException |
userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider) |
static Throwable |
valueIsNullError(int index) |
static Throwable |
writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts,
int maxDynamicPartitions,
String maxDynamicPartitionsKey) |
static org.apache.spark.SparkUnsupportedOperationException |
writeUnsupportedForBinaryFileDataSourceError() |
static Throwable |
writingJobFailedError(Throwable cause) |
public static Throwable cannotEvaluateExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotGenerateCodeForExpressionError(org.apache.spark.sql.catalyst.expressions.Expression expression)
public static Throwable cannotTerminateGeneratorError(org.apache.spark.sql.catalyst.analysis.UnresolvedGenerator generator)
public static ArithmeticException castingCauseOverflowError(Object t, DataType from, DataType to)
public static ArithmeticException castingCauseOverflowErrorInTableInsert(DataType from, DataType to, String columnName)
public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static Throwable invalidInputInCastToDatetimeError(Object value, DataType from, DataType to, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException invalidInputSyntaxForBooleanError(org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkIllegalArgumentException invalidInputInConversionError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.unsafe.types.UTF8String fmt, String hint)
public static Throwable cannotCastFromNullTypeError(DataType to)
public static Throwable cannotParseDecimalError()
public static Throwable dataTypeUnsupportedError(String dataType, String failure)
public static Throwable failedExecuteUserDefinedFunctionError(String funcCls, String inputTypes, String outputType, Throwable e)
public static ArithmeticException divideByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException intervalDividedByZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArrayIndexOutOfBoundsException invalidArrayIndexError(int index, int numElements, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArrayIndexOutOfBoundsException invalidElementAtIndexError(int index, int numElements, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static java.time.DateTimeException invalidFractionOfSecondError()
public static org.apache.spark.SparkDateTimeException ansiDateTimeParseError(Exception e)
public static org.apache.spark.SparkDateTimeException ansiDateTimeError(Exception e)
public static org.apache.spark.SparkIllegalArgumentException ansiIllegalArgumentError(String message)
public static IllegalArgumentException ansiIllegalArgumentError(IllegalArgumentException e)
public static ArithmeticException overflowInSumOfDecimalError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException overflowInIntegralDivideError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException overflowInConvError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException mapSizeExceedArraySizeWhenZipMapError(int size)
public static RuntimeException literalTypeUnsupportedError(Object v)
public static RuntimeException pivotColumnUnsupportedError(Object v, DataType dataType)
public static org.apache.spark.SparkRuntimeException noDefaultForDataTypeError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException orderedOperationUnsupportedByDataTypeError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException regexGroupIndexLessThanZeroError()
public static org.apache.spark.SparkIllegalArgumentException regexGroupIndexExceedGroupCountError(int groupCount, int groupIndex)
public static org.apache.spark.SparkIllegalArgumentException invalidUrlError(org.apache.spark.unsafe.types.UTF8String url, java.net.URISyntaxException e)
public static Throwable illegalUrlError(org.apache.spark.unsafe.types.UTF8String url)
public static org.apache.spark.SparkUnsupportedOperationException mergeUnsupportedByWindowFunctionError()
public static org.apache.spark.SparkUnsupportedOperationException dataTypeUnexpectedError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException typeUnsupportedError(DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException negativeValueUnexpectedError(org.apache.spark.sql.catalyst.expressions.Expression frequencyExpression)
public static org.apache.spark.SparkIllegalArgumentException addNewFunctionMismatchedWithFunctionError(String funcName)
public static org.apache.spark.SparkIllegalArgumentException cannotGenerateCodeForIncomparableTypeError(String codeType, DataType dataType)
public static org.apache.spark.SparkIllegalArgumentException cannotInterpolateClassIntoCodeBlockError(Object arg)
public static org.apache.spark.SparkUnsupportedOperationException customCollectionClsNotResolvedError()
public static org.apache.spark.SparkRuntimeException classUnsupportedByMapObjectsError(Class<?> cls)
public static org.apache.spark.SparkRuntimeException nullAsMapKeyNotAllowedError()
public static Throwable methodNotDeclaredError(String name)
public static org.apache.spark.SparkRuntimeException constructorNotFoundError(String cls)
public static org.apache.spark.SparkRuntimeException primaryConstructorNotFoundError(Class<?> cls)
public static org.apache.spark.SparkRuntimeException unsupportedNaturalJoinTypeError(org.apache.spark.sql.catalyst.plans.JoinType joinType)
public static org.apache.spark.SparkRuntimeException notExpectedUnresolvedEncoderError(org.apache.spark.sql.catalyst.expressions.AttributeReference attr)
public static org.apache.spark.SparkRuntimeException unsupportedEncoderError()
public static org.apache.spark.SparkRuntimeException notOverrideExpectedMethodsError(String className, String m1, String m2)
public static org.apache.spark.SparkRuntimeException failToConvertValueToJsonError(Object value, Class<?> cls, DataType dataType)
public static org.apache.spark.SparkRuntimeException unexpectedOperatorInCorrelatedSubquery(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan op, String pos)
public static org.apache.spark.SparkRuntimeException unreachableError(String err)
public static org.apache.spark.SparkRuntimeException unsupportedRoundingMode(scala.Enumeration.Value roundMode)
public static org.apache.spark.SparkRuntimeException resolveCannotHandleNestedSchema(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static org.apache.spark.SparkRuntimeException inputExternalRowCannotBeNullError()
public static String fieldCannotBeNullMsg(int index, String fieldName)
public static org.apache.spark.SparkRuntimeException fieldCannotBeNullError(int index, String fieldName)
public static Throwable unableToCreateDatabaseAsFailedToCreateDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToDropDatabaseAsFailedToDeleteDirectoryError(org.apache.spark.sql.catalyst.catalog.CatalogDatabase dbDefinition, java.io.IOException e)
public static Throwable unableToCreateTableAsFailedToCreateDirectoryError(String table, org.apache.hadoop.fs.Path defaultTableLocation, java.io.IOException e)
public static Throwable unableToDeletePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToDropTableAsFailedToDeleteDirectoryError(String table, org.apache.hadoop.fs.Path dir, java.io.IOException e)
public static Throwable unableToRenameTableAsFailedToRenameDirectoryError(String oldName, String newName, org.apache.hadoop.fs.Path oldDir, java.io.IOException e)
public static Throwable unableToCreatePartitionPathError(org.apache.hadoop.fs.Path partitionPath, java.io.IOException e)
public static Throwable unableToRenamePartitionPathError(org.apache.hadoop.fs.Path oldPartPath, java.io.IOException e)
public static org.apache.spark.SparkUnsupportedOperationException methodNotImplementedError(String methodName)
public static org.apache.spark.SparkArithmeticException arithmeticOverflowError(ArithmeticException e)
public static ArithmeticException arithmeticOverflowError(String message, String hint, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkArithmeticException unaryMinusCauseOverflowError(int originValue)
public static org.apache.spark.SparkArithmeticException binaryArithmeticCauseOverflowError(short eval1, String symbol, short eval2)
public static ArithmeticException intervalArithmeticOverflowError(String message, String hint, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static String failedToCompileMsg(Exception e)
public static Throwable internalCompilerError(org.codehaus.commons.compiler.InternalCompilerException e)
public static Throwable compilerError(org.codehaus.commons.compiler.CompileException e)
public static Throwable unsupportedTableChangeError(IllegalArgumentException e)
public static Throwable notADatasourceRDDPartitionError(Partition split)
public static org.apache.spark.SparkIllegalArgumentException dataPathNotSpecifiedError()
public static org.apache.spark.SparkIllegalArgumentException createStreamingSourceNotSpecifySchemaError()
public static org.apache.spark.SparkUnsupportedOperationException streamedOperatorUnsupportedByDataSourceError(String className, String operator)
public static org.apache.spark.SparkIllegalArgumentException multiplePathsSpecifiedError(scala.collection.Seq<String> allPaths)
public static org.apache.spark.SparkClassNotFoundException dataSourceNotFoundError(String provider, Throwable error)
public static org.apache.spark.SparkClassNotFoundException removedClassInSpark2Error(String className, Throwable e)
public static Throwable incompatibleDataSourceRegisterError(Throwable e)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInReadingDatesError(String format, String config, String option)
public static org.apache.spark.SparkUpgradeException sparkUpgradeInWritingDatesError(String format, String config)
public static org.apache.spark.SparkUnsupportedOperationException buildReaderUnsupportedForFileFormatError(String format)
public static Throwable taskFailedWhileWritingRowsError(String path, Throwable cause)
public static org.apache.spark.SparkFileNotFoundException readCurrentFileNotFoundError(java.io.FileNotFoundException e)
public static Throwable saveModeUnsupportedError(Object saveMode, boolean pathExists)
public static Throwable cannotClearOutputDirectoryError(org.apache.hadoop.fs.Path staticPrefixPath)
public static Throwable cannotClearPartitionDirectoryError(org.apache.hadoop.fs.Path path)
public static org.apache.spark.SparkRuntimeException failedToCastValueToDataTypeForPartitionColumnError(String value, DataType dataType, String columnName)
public static Throwable endOfStreamError()
public static org.apache.spark.SparkIllegalArgumentException fallbackV1RelationReportsInconsistentSchemaError(StructType v2Schema, StructType v1Schema)
public static Throwable noRecordsFromEmptyDataReaderError()
public static org.apache.spark.SparkFileNotFoundException fileNotFoundError(java.io.FileNotFoundException e)
public static Throwable unsupportedSchemaColumnConvertError(String filePath, String column, String logicalType, String physicalType, Exception e)
public static Throwable cannotReadFilesError(Throwable e, String path)
public static Throwable cannotCreateColumnarReaderError()
public static org.apache.spark.SparkIllegalArgumentException invalidNamespaceNameError(String[] namespace)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedPartitionTransformError(Transform transform)
public static org.apache.spark.SparkIllegalArgumentException missingDatabaseLocationError()
public static org.apache.spark.SparkUnsupportedOperationException cannotRemoveReservedPropertyError(String property)
public static Throwable writingJobFailedError(Throwable cause)
public static Throwable commitDeniedError(int partId, long taskId, int attemptId, int stageId, int stageAttempt)
public static Throwable unsupportedTableWritesError(Identifier ident)
public static org.apache.spark.SparkUnsupportedOperationException cannotCreateJDBCTableWithPartitionsError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedUserSpecifiedSchemaError()
public static org.apache.spark.SparkUnsupportedOperationException writeUnsupportedForBinaryFileDataSourceError()
public static Throwable fileLengthExceedsMaxLengthError(org.apache.hadoop.fs.FileStatus status, int maxLength)
public static org.apache.spark.SparkRuntimeException unsupportedFieldNameError(String fieldName)
public static org.apache.spark.SparkIllegalArgumentException cannotSpecifyBothJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static org.apache.spark.SparkIllegalArgumentException missingJdbcTableNameAndQueryError(String jdbcTableName, String jdbcQueryString)
public static org.apache.spark.SparkIllegalArgumentException emptyOptionError(String optionName)
public static org.apache.spark.SparkIllegalArgumentException invalidJdbcTxnIsolationLevelError(String jdbcTxnIsolationLevel, String value)
public static org.apache.spark.SparkIllegalArgumentException cannotGetJdbcTypeError(DataType dt)
public static Throwable unrecognizedSqlTypeError(int sqlType)
public static org.apache.spark.SparkSQLException unsupportedJdbcTypeError(String content)
public static org.apache.spark.SparkIllegalArgumentException unsupportedArrayElementTypeBasedOnBinaryError(DataType dt)
public static org.apache.spark.SparkIllegalArgumentException nestedArraysUnsupportedError()
public static org.apache.spark.SparkIllegalArgumentException cannotTranslateNonNullValueForFieldError(int pos)
public static org.apache.spark.SparkIllegalArgumentException invalidJdbcNumPartitionsError(int n, String jdbcNumPartitions)
public static Throwable multiActionAlterError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException dataTypeUnsupportedYetError(DataType dataType)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationForDataTypeError(DataType dataType)
public static Throwable inputFilterNotFullyConvertibleError(String owner)
public static Throwable cannotReadFooterForFileError(org.apache.hadoop.fs.Path file, Exception e)
public static org.apache.spark.SparkRuntimeException foundDuplicateFieldInCaseInsensitiveModeError(String requiredFieldName, String matchedOrcFields)
public static org.apache.spark.SparkRuntimeException foundDuplicateFieldInFieldIdLookupModeError(int requiredId, String matchedFields)
public static Throwable failedToMergeIncompatibleSchemasError(StructType left, StructType right, Throwable e)
public static org.apache.spark.SparkUnsupportedOperationException ddlUnsupportedTemporarilyError(String ddl)
public static Throwable executeBroadcastTimeoutError(long timeout, scala.Option<java.util.concurrent.TimeoutException> ex)
public static org.apache.spark.SparkIllegalArgumentException cannotCompareCostWithTargetCostError(String cost)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedArrowTypeError(org.apache.arrow.vector.types.pojo.ArrowType typeName)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedDataTypeError(DataType typeName)
public static Throwable notSupportTypeError(DataType dataType)
public static org.apache.spark.SparkRuntimeException notSupportNonPrimitiveTypeError()
public static Throwable unsupportedTypeError(DataType dataType)
public static Throwable useDictionaryEncodingWhenDictionaryOverflowError()
public static Throwable endOfIteratorError()
public static Throwable cannotAllocateMemoryToGrowBytesToBytesMapError()
public static Throwable cannotAcquireMemoryToBuildLongHashedRelationError(long size, long got)
public static Throwable cannotAcquireMemoryToBuildUnsafeHashedRelationError()
public static org.apache.spark.SparkUnsupportedOperationException rowLargerThan256MUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationWithUniqueKeysExceededError()
public static org.apache.spark.SparkUnsupportedOperationException cannotBuildHashedRelationLargerThan8GError()
public static Throwable failedToPushRowIntoRowQueueError(String rowQueue)
public static org.apache.spark.SparkRuntimeException unexpectedWindowFunctionFrameError(String frame)
public static org.apache.spark.SparkIllegalArgumentException cannotParseStatisticAsPercentileError(String stats, NumberFormatException e)
public static org.apache.spark.SparkIllegalArgumentException statisticNotRecognizedError(String stats)
public static org.apache.spark.SparkIllegalArgumentException unknownColumnError(String unknownColumn)
public static org.apache.spark.SparkIllegalArgumentException unexpectedAccumulableUpdateValueError(Object o)
public static ArithmeticException unscaledValueTooLargeForPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkArithmeticException decimalPrecisionExceedsMaxPrecisionError(int precision, int maxPrecision)
public static org.apache.spark.SparkArithmeticException outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str)
public static org.apache.spark.SparkRuntimeException unsupportedArrayTypeError(Class<?> clazz)
public static org.apache.spark.SparkRuntimeException unsupportedJavaTypeError(Class<?> clazz)
public static org.apache.spark.SparkRuntimeException failedParsingStructTypeError(String raw)
public static Throwable cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale, int rightScale)
public static Throwable cannotMergeIncompatibleDataTypesError(DataType left, DataType right)
public static org.apache.spark.SparkRuntimeException exceedMapSizeLimitError(int size)
public static org.apache.spark.SparkRuntimeException duplicateMapKeyFoundError(Object key)
public static org.apache.spark.SparkRuntimeException mapDataKeyArrayLengthDiffersFromValueArrayLengthError()
public static org.apache.spark.SparkDateTimeException fieldDiffersFromDerivedLocalDateError(java.time.temporal.ChronoField field, int actual, int expected, java.time.LocalDate candidate)
public static Throwable failToParseDateTimeInNewParserError(String s, Throwable e)
public static Throwable failToRecognizePatternAfterUpgradeError(String pattern, Throwable e, String docroot)
public static org.apache.spark.SparkRuntimeException failToRecognizePatternError(String pattern, Throwable e, String docroot)
public static Throwable registeringStreamingQueryListenerError(Exception e)
public static Throwable concurrentQueryInstanceError()
public static org.apache.spark.SparkRuntimeException cannotParseJsonArraysAsStructsError()
public static org.apache.spark.SparkRuntimeException cannotParseStringAsDataTypeError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken token, DataType dataType)
public static org.apache.spark.SparkRuntimeException cannotParseStringAsDataTypeError(String pattern, String value, DataType dataType)
public static org.apache.spark.SparkRuntimeException emptyJsonFieldValueError(DataType dataType)
public static org.apache.spark.SparkRuntimeException cannotParseJSONFieldError(com.fasterxml.jackson.core.JsonParser parser, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType)
public static org.apache.spark.SparkRuntimeException cannotParseJSONFieldError(String fieldName, String fieldValue, com.fasterxml.jackson.core.JsonToken jsonType, DataType dataType)
public static org.apache.spark.SparkRuntimeException rootConverterReturnNullError()
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInBeanClassError(Class<?> clazz)
public static org.apache.spark.SparkUnsupportedOperationException cannotHaveCircularReferencesInClassError(String t)
public static org.apache.spark.SparkUnsupportedOperationException cannotUseInvalidJavaIdentifierAsFieldNameError(String fieldName, org.apache.spark.sql.catalyst.WalkedTypePath walkedTypePath)
public static org.apache.spark.SparkUnsupportedOperationException cannotFindEncoderForTypeError(String typeName, String docroot)
public static org.apache.spark.SparkUnsupportedOperationException attributesForTypeUnsupportedError(org.apache.spark.sql.catalyst.ScalaReflection.Schema schema)
public static org.apache.spark.SparkUnsupportedOperationException cannotFindConstructorForTypeError(String tpe)
public static org.apache.spark.SparkRuntimeException paramExceedOneCharError(String paramName)
public static org.apache.spark.SparkRuntimeException paramIsNotIntegerError(String paramName, String value)
public static Throwable paramIsNotBooleanValueError(String paramName)
public static org.apache.spark.SparkRuntimeException foundNullValueForNotNullableFieldError(String name)
public static org.apache.spark.SparkRuntimeException malformedCSVRecordError(String badRecord)
public static org.apache.spark.SparkUnsupportedOperationException elementsOfTupleExceedLimitError()
public static org.apache.spark.SparkRuntimeException expressionDecodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static org.apache.spark.SparkRuntimeException expressionEncodingError(Exception e, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> expressions)
public static org.apache.spark.SparkRuntimeException classHasUnexpectedSerializerError(String clsName, org.apache.spark.sql.catalyst.expressions.Expression objSerializer)
public static org.apache.spark.SparkRuntimeException cannotGetOuterPointerForInnerClassError(Class<?> innerCls)
public static Throwable userDefinedTypeNotAnnotatedAndRegisteredError(UserDefinedType<?> udt)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperandTypeForSizeFunctionError(DataType dataType)
public static org.apache.spark.SparkRuntimeException unexpectedValueForStartInFunctionError(String prettyName)
public static org.apache.spark.SparkRuntimeException unexpectedValueForLengthInFunctionError(String prettyName)
public static RuntimeException invalidIndexOfZeroError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkRuntimeException concatArraysWithElementsExceedLimitError(long numberOfElements)
public static org.apache.spark.SparkRuntimeException flattenArraysWithElementsExceedLimitError(long numberOfElements)
public static org.apache.spark.SparkRuntimeException createArrayWithElementsExceedLimitError(Object count)
public static org.apache.spark.SparkRuntimeException unionArrayWithElementsExceedLimitError(int length)
public static org.apache.spark.SparkUnsupportedOperationException initialTypeNotTargetDataTypeError(DataType dataType, String target)
public static org.apache.spark.SparkUnsupportedOperationException initialTypeNotTargetDataTypesError(DataType dataType)
public static Throwable malformedRecordsDetectedInSchemaInferenceError(Throwable e)
public static Throwable malformedJSONError()
public static Throwable malformedRecordsDetectedInSchemaInferenceError(DataType dataType)
public static org.apache.spark.SparkUnsupportedOperationException decorrelateInnerQueryThroughPlanUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan)
public static org.apache.spark.SparkRuntimeException methodCalledInAnalyzerNotAllowedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSafelyMergeSerdePropertiesError(scala.collection.immutable.Map<String,String> props1, scala.collection.immutable.Map<String,String> props2, scala.collection.immutable.Set<String> conflictKeys)
public static org.apache.spark.SparkUnsupportedOperationException pairUnsupportedAtFunctionError(org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r1, org.apache.spark.sql.catalyst.plans.logical.statsEstimation.ValueInterval r2, String function)
public static <TreeType extends org.apache.spark.sql.catalyst.trees.TreeNode<?>> org.apache.spark.SparkRuntimeException onceStrategyIdempotenceIsBrokenForBatchError(String batchName, TreeType plan, TreeType reOptimized)
public static Throwable ruleIdNotFoundForRuleError(String ruleName)
public static org.apache.spark.SparkRuntimeException cannotCreateArrayWithElementsExceedLimitError(long numElements, String additionalErrorMessage)
public static Throwable malformedRecordsDetectedInRecordParsingError(String badRecord, org.apache.spark.sql.catalyst.util.BadRecordException e)
public static org.apache.spark.SparkRuntimeException remoteOperationsUnsupportedError()
public static Throwable invalidKerberosConfigForHiveServer2Error()
public static Throwable parentSparkUIToAttachTabNotFoundError()
public static org.apache.spark.SparkUnsupportedOperationException inferSchemaUnsupportedForHiveError()
public static Throwable requestedPartitionsMismatchTablePartitionsError(org.apache.spark.sql.catalyst.catalog.CatalogTable table, scala.collection.immutable.Map<String,scala.Option<String>> partition)
public static Throwable dynamicPartitionKeyNotAmongWrittenPartitionPathsError(String key)
public static org.apache.spark.SparkRuntimeException cannotRemovePartitionDirError(org.apache.hadoop.fs.Path partitionPath)
public static org.apache.spark.SparkRuntimeException cannotCreateStagingDirError(String message, java.io.IOException e)
public static org.apache.spark.SparkClassNotFoundException serDeInterfaceNotFoundError(NoClassDefFoundError e)
public static Throwable convertHiveTableToCatalogTableError(SparkException e, String dbName, String tableName)
public static Throwable cannotRecognizeHiveTypeError(org.apache.spark.sql.catalyst.parser.ParseException e, String fieldType, String fieldName)
public static org.apache.spark.SparkUnsupportedOperationException getTablesByTypeUnsupportedByHiveVersionError()
public static org.apache.spark.SparkUnsupportedOperationException dropTableWithPurgeUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException alterTableWithDropPartitionAndPurgeUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException invalidPartitionFilterError()
public static org.apache.spark.SparkRuntimeException getPartitionMetadataByFilterError(java.lang.reflect.InvocationTargetException e)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedHiveMetastoreVersionError(String version, String key)
public static org.apache.spark.SparkClassNotFoundException loadHiveClientCausesNoClassDefFoundError(NoClassDefFoundError cnf, scala.collection.Seq<java.net.URL> execJars, String key, java.lang.reflect.InvocationTargetException e)
public static Throwable cannotFetchTablesOfDatabaseError(String dbName, Exception e)
public static Throwable illegalLocationClauseForViewPartitionError()
public static Throwable renamePathAsExistsPathError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static org.apache.spark.SparkFileAlreadyExistsException renameAsExistsPathError(org.apache.hadoop.fs.Path dstPath)
public static Throwable renameSrcPathNotFoundError(org.apache.hadoop.fs.Path srcPath)
public static Throwable failedRenameTempFileError(org.apache.hadoop.fs.Path srcPath, org.apache.hadoop.fs.Path dstPath)
public static Throwable legacyMetadataPathExistsError(org.apache.hadoop.fs.Path metadataPath, org.apache.hadoop.fs.Path legacyMetadataPath)
public static org.apache.spark.SparkRuntimeException partitionColumnNotFoundInSchemaError(String col, StructType schema)
public static Throwable stateNotDefinedOrAlreadyRemovedError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutDurationError()
public static org.apache.spark.SparkUnsupportedOperationException cannotGetEventTimeWatermarkError()
public static org.apache.spark.SparkUnsupportedOperationException cannotSetTimeoutTimestampError()
public static org.apache.spark.SparkFileNotFoundException batchMetadataFileNotFoundError(org.apache.hadoop.fs.Path batchMetadataFile)
public static org.apache.spark.SparkConcurrentModificationException multiStreamingQueriesUsingPathConcurrentlyError(String path, org.apache.hadoop.fs.FileAlreadyExistsException e)
public static org.apache.spark.SparkUnsupportedOperationException addFilesWithAbsolutePathUnsupportedError(String commitProtocol)
public static org.apache.spark.SparkUnsupportedOperationException microBatchUnsupportedByDataSourceError(String srcName, String disabledSources, Table table)
public static org.apache.spark.SparkUnsupportedOperationException cannotExecuteStreamingRelationExecError()
public static org.apache.spark.SparkUnsupportedOperationException invalidStreamingOutputModeError(scala.Option<OutputMode> outputMode)
public static Throwable invalidCatalogNameError(String name)
public static Throwable catalogPluginClassNotFoundError(String name)
public static Throwable catalogPluginClassNotImplementedError(String name, String pluginClassName)
public static Throwable catalogPluginClassNotFoundForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToFindPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable catalogFailToCallPublicNoArgConstructorError(String name, String pluginClassName, Exception e)
public static Throwable cannotInstantiateAbstractCatalogPluginClassError(String name, String pluginClassName, Exception e)
public static Throwable failedToInstantiateConstructorForCatalogError(String name, String pluginClassName, Exception e)
public static Throwable noSuchElementExceptionError()
public static Throwable noSuchElementExceptionError(String key)
public static org.apache.spark.SparkUnsupportedOperationException cannotMutateReadOnlySQLConfError()
public static org.apache.spark.SparkUnsupportedOperationException cannotCloneOrCopyReadOnlySQLConfError()
public static org.apache.spark.SparkRuntimeException cannotGetSQLConfInSchedulerEventLoopThreadError()
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationExceptionError()
public static org.apache.spark.SparkUnsupportedOperationException nullLiteralsCannotBeCastedError(String name)
public static Throwable notUserDefinedTypeError(String name, String userClass)
public static Throwable cannotLoadUserDefinedTypeError(String name, String userClass)
public static org.apache.spark.SparkUnsupportedOperationException notPublicClassError(String name)
public static org.apache.spark.SparkUnsupportedOperationException primitiveTypesNotSupportedError()
public static org.apache.spark.SparkUnsupportedOperationException fieldIndexOnRowWithoutSchemaError()
public static Throwable valueIsNullError(int index)
public static Throwable onlySupportDataSourcesProvidingFileFormatError(String providingClass)
public static Throwable cannotRestorePermissionsForPathError(org.apache.hadoop.fs.permission.FsPermission permission, org.apache.hadoop.fs.Path path)
public static org.apache.spark.SparkSecurityException failToSetOriginalACLBackError(String aclEntries, org.apache.hadoop.fs.Path path, Throwable e)
public static Throwable multiFailuresInStageMaterializationError(Throwable error)
public static org.apache.spark.SparkUnsupportedOperationException unrecognizedCompressionSchemaTypeIDError(int typeId)
public static org.apache.spark.SparkSQLFeatureNotSupportedException getParentLoggerNotImplementedError(String className)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForTypeError(DecimalType t, String parquetType)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForDecimalTypeError(DecimalType t, String parquetType)
public static org.apache.spark.SparkRuntimeException cannotCreateParquetConverterForDataTypeError(DataType t, String parquetType)
public static org.apache.spark.SparkUnsupportedOperationException cannotAddMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException userSpecifiedSchemaUnsupportedByDataSourceError(TableProvider provider)
public static org.apache.spark.SparkUnsupportedOperationException cannotDropMultiPartitionsOnNonatomicPartitionTableError(String tableName)
public static org.apache.spark.SparkUnsupportedOperationException truncateMultiPartitionUnsupportedError(String tableName)
public static Throwable overwriteTableByUnsupportedExpressionError(Table table)
public static Throwable dynamicPartitionOverwriteUnsupportedByTableError(Table table)
public static Throwable failedMergingSchemaError(StructType schema, SparkException e)
public static Throwable cannotBroadcastTableOverMaxTableRowsError(long maxBroadcastTableRows, long numRows)
public static Throwable cannotBroadcastTableOverMaxTableBytesError(long maxBroadcastTableBytes, long dataSize)
public static Throwable notEnoughMemoryToBuildAndBroadcastTableError(OutOfMemoryError oe, scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> tables)
public static org.apache.spark.SparkUnsupportedOperationException executeCodePathUnsupportedError(String execName)
public static org.apache.spark.SparkUnsupportedOperationException cannotMergeClassWithOtherClassError(String className, String otherClass)
public static org.apache.spark.SparkUnsupportedOperationException continuousProcessingUnsupportedByDataSourceError(String sourceName)
public static Throwable failedToReadDataError(Throwable failureReason)
public static Throwable failedToGenerateEpochMarkerError(Throwable failureReason)
public static Throwable foreachWriterAbortedDueToTaskFailureError()
public static Throwable incorrectRampUpRate(long rowsPerSecond, long maxSeconds, long rampUpTimeSeconds)
public static Throwable incorrectEndOffset(long rowsPerSecond, long maxSeconds, long endSeconds)
public static Throwable failedToReadDeltaFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, int keySize)
public static Throwable failedToReadSnapshotFileError(org.apache.hadoop.fs.Path fileToRead, String clazz, String message)
public static org.apache.spark.SparkUnsupportedOperationException cannotPurgeAsBreakInternalStateError()
public static org.apache.spark.SparkUnsupportedOperationException cleanUpSourceFilesUnsupportedError()
public static org.apache.spark.SparkUnsupportedOperationException latestOffsetNotCalledError()
public static Throwable legacyCheckpointDirectoryExistsError(org.apache.hadoop.fs.Path checkpointPath, String legacyCheckpointDir)
public static Throwable subprocessExitedError(int exitCode, org.apache.spark.util.CircularBuffer stderrBuffer, Throwable cause)
public static Throwable outputDataTypeUnsupportedByNodeWithoutSerdeError(String nodeName, DataType dt)
public static org.apache.spark.SparkArrayIndexOutOfBoundsException invalidStartIndexError(int numRows, int startIndex)
public static org.apache.spark.SparkConcurrentModificationException concurrentModificationOnExternalAppendOnlyUnsafeRowArrayError(String className)
public static org.apache.spark.SparkUnsupportedOperationException doExecuteBroadcastNotImplementedError(String nodeName)
public static Throwable defaultDatabaseNotExistsError(String defaultDatabase)
public static Throwable databaseNameConflictWithSystemPreservedDatabaseError(String globalTempDB)
public static org.apache.spark.SparkSQLFeatureNotSupportedException commentOnTableUnsupportedError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedUpdateColumnNullabilityError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException renameColumnUnsupportedForOlderMySQLError()
public static SparkException failedToExecuteQueryError(Throwable e)
public static org.apache.spark.SparkUnsupportedOperationException nestedFieldUnsupportedError(String colName)
public static Throwable transformationsAndActionsNotInvokedByDriverError()
public static Throwable repeatedPivotsUnsupportedError(String clause, String operation)
public static Throwable pivotNotAfterGroupByUnsupportedError()
public static RuntimeException invalidAesKeyLengthError(int actualLength)
public static RuntimeException aesModeUnsupportedError(String mode, String padding)
public static RuntimeException aesCryptoError(String detailMessage)
public static org.apache.spark.SparkUnsupportedOperationException hiveTableWithAnsiIntervalsError(String tableName)
public static Throwable cannotConvertOrcTimestampToTimestampNTZError()
public static Throwable cannotConvertOrcTimestampNTZToTimestampLTZError()
public static Throwable writePartitionExceedConfigSizeWhenDynamicPartitionError(int numWrittenParts, int maxDynamicPartitions, String maxDynamicPartitionsKey)
public static org.apache.spark.SparkIllegalArgumentException invalidNumberFormatError(String valueType, String input, String format)
public static org.apache.spark.SparkUnsupportedOperationException unsupportedMultipleBucketTransformsError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedCreateNamespaceCommentError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedRemoveNamespaceCommentError()
public static org.apache.spark.SparkSQLFeatureNotSupportedException unsupportedDropNamespaceRestrictError()
public static ArithmeticException timestampAddOverflowError(long micros, int amount, String unit)
public static Throwable invalidBucketFile(String path)
public static Throwable multipleRowSubqueryError(org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static Throwable comparatorReturnsNull(String firstValue, String secondValue)
public static RuntimeException invalidPatternError(String funcName, String pattern, Throwable cause)
public static org.apache.spark.SparkIllegalArgumentException tooManyArrayElementsError(int numElements, int elementSize)
public static org.apache.spark.SparkIllegalArgumentException invalidEmptyLocationError(String location)
public static Throwable malformedProtobufMessageDetectedInMessageParsingError(Throwable e)
public static Throwable locationAlreadyExists(org.apache.spark.sql.catalyst.TableIdentifier tableId, org.apache.hadoop.fs.Path location)
public static String toSQLValue(Object v, DataType t)
public static String toSQLStmt(String text)
public static String toSQLId(scala.collection.Seq<String> parts)
public static String toSQLId(String parts)
public static String toSQLType(org.apache.spark.sql.types.AbstractDataType t)
public static String toSQLConf(String conf)
public static String toSQLConfVal(String conf)
public static String toDSOption(String option)
public static String toSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e)
public static String toSQLSchema(String schema)
public static String getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)
public static QueryContext[] getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)