Package org.apache.spark.sql.errors
Class QueryCompilationErrors
Object
org.apache.spark.sql.errors.QueryCompilationErrors
Object for grouping error messages from exceptions thrown during query compilation.
As commands are executed eagerly, this also includes errors thrown during the execution of
commands, which users can see immediately.
-
Constructor Summary
Constructors -
Method Summary
Modifier and TypeMethodDescriptionstatic ThrowableactionNotAllowedOnTableSincePartitionMetadataNotStoredError(String action, String tableName) static ThrowableactionNotAllowedOnTableWithFilesourcePartitionManagementDisabledError(String action, String tableName) static Throwablestatic ThrowableaddNewDefaultColumnToExistingTableNotAllowed(String statementType, String dataSource) static Throwablestatic Throwablestatic Throwablestatic ThrowableaggregationFunctionAppliedOnNonNumericColumnError(String pivotColumn, int maxValues) static ThrowableaggregationNotAllowedInMergeCondition(String condName, org.apache.spark.sql.catalyst.expressions.Expression cond) static ThrowablealiasesNumberNotMatchUDTFOutputError(int aliasesSize, String aliasesNames) static ThrowablealiasNumberNotMatchColumnNumberError(int columnSize, int outputSize, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowablealterAddColNotSupportDatasourceTableError(Object tableType, org.apache.spark.sql.catalyst.TableIdentifier table) static ThrowablealterAddColNotSupportViewError(org.apache.spark.sql.catalyst.TableIdentifier table) static ThrowablealterColumnCannotFindColumnInV1TableError(String colName, org.apache.spark.sql.connector.catalog.V1Table v1Table) static Throwablestatic Throwablestatic ThrowablealterTableChangeColumnNotSupportedForColumnTypeError(String tableName, StructField originColumn, StructField newColumn) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowableambiguousAttributesInSelfJoinError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> ambiguousAttrs) static ThrowableambiguousColumnOrFieldError(scala.collection.Seq<String> name, int numMatches, org.apache.spark.sql.catalyst.trees.Origin context) static ThrowableambiguousLateralColumnAliasError(String name, int numOfMatches) static ThrowableambiguousLateralColumnAliasError(scala.collection.Seq<String> nameParts, int numOfMatches) static ThrowableambiguousReferenceError(String name, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> ambiguousReferences) static ThrowableambiguousReferenceToFieldsError(String field, int numberOfAppearance) static Throwablestatic Throwablestatic Throwablestatic ThrowableanalyzingColumnStatisticsNotSupportedForColumnTypeError(String name, DataType dataType) static Throwablestatic Throwablestatic ThrowableavroIncompatibleReadError(String avroPath, String sqlPath, String avroType, String sqlType) static ThrowablebatchWriteCapabilityError(Table table, String v2WriteClassName, String v1WriteClassName) static Throwablestatic ThrowablebucketByUnsupportedByOperationError(String operation) static ThrowablebucketingColumnCannotBePartOfPartitionColumnsError(String bucketCol, scala.collection.Seq<String> normalizedPartCols) static ThrowablebucketSortingColumnCannotBePartOfPartitionColumnsError(String sortCol, scala.collection.Seq<String> normalizedPartCols) static Throwablestatic Throwablestatic ThrowablecannotConvertCatalystTypeToProtobufTypeError(scala.collection.Seq<String> sqlColumn, String protobufColumn, DataType sqlType, String protobufType) static Throwablestatic ThrowablecannotConvertProtobufTypeToCatalystTypeError(String protobufType, DataType sqlType, Throwable cause) static ThrowablecannotConvertProtobufTypeToSqlTypeError(String protobufColumn, scala.collection.Seq<String> sqlColumn, String protobufType, DataType sqlType) static ThrowablecannotConvertSqlTypeToProtobufError(String protobufType, DataType sqlType, Throwable cause) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowablecannotCreateTableWithBothProviderAndSerdeError(scala.Option<String> provider, scala.Option<org.apache.spark.sql.catalyst.plans.logical.SerdeInfo> maybeSerdeInfo) static Throwablestatic ThrowablecannotCreateViewNotEnoughColumnsError(org.apache.spark.sql.catalyst.TableIdentifier viewIdent, scala.collection.Seq<String> expected, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query) static ThrowablecannotCreateViewTooManyColumnsError(org.apache.spark.sql.catalyst.TableIdentifier viewIdent, scala.collection.Seq<String> expected, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query) static ThrowablecannotDeleteTableWhereFiltersError(Table table, Predicate[] filters) static ThrowablecannotDropBuiltinFuncError(String functionName) static ThrowablecannotDropDefaultDatabaseError(String database) static Throwablestatic ThrowablecannotDropNonemptyNamespaceError(scala.collection.Seq<String> namespace) static Throwablestatic ThrowablecannotFindCatalystTypeInProtobufSchemaError(String catalystFieldPath) static ThrowablecannotFindColumnError(String name, String[] fieldNames) static ThrowablecannotFindColumnInRelationOutputError(String colName, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan relation) static AnalysisExceptioncannotFindDescriptorFileError(String filePath, Throwable cause) static ThrowablecannotFindPartitionColumnInPartitionSchemaError(StructField readField, StructType partitionSchema) static Throwablestatic ThrowablecannotLoadClassNotOnClassPathError(String className) static ThrowablecannotLoadClassWhenRegisteringFunctionError(String className, org.apache.spark.sql.catalyst.FunctionIdentifier func) static ThrowablecannotModifyValueOfSparkConfigError(String key, String docroot) static Throwablestatic ThrowablecannotOperateOnHiveDataSourceFilesError(String operation) static Throwablestatic ThrowablecannotOverwriteTableThatIsBeingReadFromError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) static ThrowablecannotParseIntervalError(String delayThreshold, Throwable e) static Throwablestatic ThrowablecannotRefreshBuiltInFuncError(String functionName) static ThrowablecannotRefreshTempFuncError(String functionName) static Throwablestatic Throwablestatic ThrowablecannotRenameTempViewToExistingTableError(org.apache.spark.sql.catalyst.TableIdentifier newName) static ThrowablecannotRenameTempViewWithDatabaseSpecifiedError(org.apache.spark.sql.catalyst.TableIdentifier oldName, org.apache.spark.sql.catalyst.TableIdentifier newName) static ThrowablecannotReplaceMissingTableError(Identifier tableIdentifier) static ThrowablecannotReplaceMissingTableError(Identifier tableIdentifier, scala.Option<Throwable> cause) static ThrowablecannotResolveAttributeError(String name, String outputStr) static ThrowablecannotResolveColumnGivenInputColumnsError(String col, String inputColumns) static ThrowablecannotResolveColumnNameAmongAttributesError(String colName, String fieldNames) static ThrowablecannotResolveStarExpandGivenInputColumnsError(String targetString, String columns) static ThrowablecannotRetrieveTableOrViewNotInSameDatabaseError(scala.collection.Seq<org.apache.spark.sql.catalyst.QualifiedTableName> qualifiedTableNames) static Throwablestatic Throwablestatic ThrowablecannotSpecifyDatabaseForTempViewError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) static ThrowablecannotSpecifyWindowFrameError(String prettyName) static ThrowablecannotTranslateExpressionToSourceFilterError(org.apache.spark.sql.catalyst.expressions.Expression f) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowablecannotWriteNotEnoughColumnsToTableError(String tableName, scala.collection.Seq<String> expected, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> queryOutput) static ThrowablecannotWriteTooManyColumnsToTableError(String tableName, scala.collection.Seq<String> expected, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> queryOutput) static ThrowablecatalogOperationNotSupported(CatalogPlugin catalog, String operation) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowablecmdOnlyWorksOnPartitionedTablesError(String operation, String tableIdentWithDB) static ThrowablecmdOnlyWorksOnTableWithLocationError(String cmd, String tableIdentWithDB) static ThrowablecolumnAlreadyExistsError(String columnName) static ThrowablecolumnNotDefinedInTableError(String colType, String colName, String tableName, scala.collection.Seq<String> tableCols) static AnalysisExceptioncolumnNotFoundError(String colName) static ThrowablecolumnNotFoundInExistingColumnsError(String columnType, String columnName, scala.collection.Seq<String> validColumnNames) static ThrowablecolumnNotFoundInSchemaError(StructField col, scala.Option<StructType> tableSchema) static ThrowablecolumnNotInGroupByClauseError(org.apache.spark.sql.catalyst.expressions.Expression expression) static ThrowablecolumnStatisticsDeserializationNotSupportedError(String name, DataType dataType) static ThrowablecolumnStatisticsSerializationNotSupportedError(String colName, DataType dataType) static ThrowablecolumnTypeNotSupportStatisticsCollectionError(String name, org.apache.spark.sql.catalyst.TableIdentifier tableIdent, DataType dataType) static Throwablestatic ThrowablecommandNotSupportNestedColumnError(String command, String quoted) static Throwablestatic ThrowableconfigRemovedInVersionError(String configName, String version, String comment) static ThrowableconflictingAttributesInJoinConditionError(org.apache.spark.sql.catalyst.expressions.AttributeSet conflictingAttrs, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan outerPlan, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan subplan) static ThrowablecorruptedTableNameContextInCatalogError(int numParts, int index) static ThrowablecorruptedViewQueryOutputColumnsInCatalogError(String numCols, int index) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowabledataPathNotExistError(String path) static ThrowabledataSchemaNotSpecifiedError(String format) static ThrowabledataSchemaNotSpecifiedError(String format, String fileCatalog) static ThrowabledataSourceOutputModeUnsupportedError(String className, OutputMode outputMode) static ThrowabledataTypeMismatchForDeserializerError(DataType dataType, String desiredType) static Throwablestatic ThrowabledataTypeUnsupportedByClassError(DataType x, String className) static ThrowabledataTypeUnsupportedByDataSourceError(String format, StructField column) static ThrowabledataTypeUnsupportedByExtractValueError(DataType dataType, org.apache.spark.sql.catalyst.expressions.Expression extraction, org.apache.spark.sql.catalyst.expressions.Expression child) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowabledefaultReferencesNotAllowedInDataSource(String statementType, String dataSource) static ThrowabledefaultValuesDataTypeError(String statement, String colName, String defaultValue, DataType expectedType, DataType actualType) static ThrowabledefaultValuesMayNotContainSubQueryExpressions(String statement, String colName, String defaultValue) static ThrowabledefaultValuesUnresolvedExprError(String statement, String colName, String defaultValue, Throwable cause) static Throwablestatic Throwablestatic Throwablestatic AnalysisExceptiondescriptorParseError(Throwable cause) static ThrowabledoubleNamedArgumentReference(String functionName, String parameterName) static ThrowabledropNonExistentColumnsNotSupportedError(scala.collection.Seq<String> nonExistentColumnNames) static Throwablestatic ThrowableemptyWindowExpressionError(org.apache.spark.sql.catalyst.plans.logical.Window expr) static ThrowableescapeCharacterAtTheEndError(String pattern) static ThrowableescapeCharacterInTheMiddleError(String pattern, String char_) static ThrowableexpectPersistentFuncError(String name, String cmd, scala.Option<String> mismatchHint, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableexpectTableNotViewError(scala.collection.Seq<String> nameParts, boolean isTemp, String cmd, scala.Option<String> mismatchHint, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableexpectTableOrPermanentViewNotTempViewError(scala.collection.Seq<String> nameParts, String cmd, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableexpectViewNotTableError(org.apache.spark.sql.catalyst.analysis.ResolvedTable v, String cmd, scala.Option<String> mismatchHint, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableexpectViewNotTempViewError(scala.collection.Seq<String> nameParts, String cmd, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableexpressionWithMultiWindowExpressionsError(org.apache.spark.sql.catalyst.expressions.NamedExpression expr, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.WindowSpecDefinition> distinctWindowSpec) static ThrowableexpressionWithoutWindowExpressionError(org.apache.spark.sql.catalyst.expressions.NamedExpression expr) static ThrowablefailedToFindAvroDataSourceError(String provider) static ThrowablefailedToFindKafkaDataSourceError(String provider) static ThrowablefailedToParseExistenceDefaultAsLiteral(String fieldName, String defaultValue) static ThrowablefailedToRebuildExpressionError(Filter filter) static ThrowablefailToResolveDataSourceForTableError(org.apache.spark.sql.catalyst.catalog.CatalogTable table, String key) static ThrowablefailToTruncateTableWhenRemovingDataError(String tableIdentWithDB, org.apache.hadoop.fs.Path path, Throwable e) static ThrowablefieldNumberMismatchForDeserializerError(StructType schema, int maxOrdinal) static ThrowablefindMultipleDataSourceError(String provider, scala.collection.Seq<String> sourceNames) static ThrowablefoundDifferentWindowFunctionTypeError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> windowExpressions) static ThrowablefoundRecursionInProtobufSchema(String fieldDescriptor) static ThrowablefuncBuildError(String funcName, Exception cause) static ThrowablefunctionAlreadyExistsError(org.apache.spark.sql.catalyst.FunctionIdentifier func) static ThrowablefunctionCannotProcessInputError(UnboundFunction unbound, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> arguments, UnsupportedOperationException unsupported) static ThrowablefunctionWithUnsupportedSyntaxError(String prettyName, String syntax) static ThrowablegeneratorNotExpectedError(org.apache.spark.sql.catalyst.FunctionIdentifier name, String classCanonicalName) static ThrowablegeneratorOutsideSelectError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) static QueryContext[]getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) static StringgetSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) static ThrowablegroupByPositionRangeError(int index, int size) static ThrowablegroupByPositionRefersToAggregateFunctionError(int index, org.apache.spark.sql.catalyst.expressions.Expression expr) static ThrowablegroupingColInvalidError(org.apache.spark.sql.catalyst.expressions.Expression groupingCol, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupByExprs) static ThrowablegroupingIDMismatchError(org.apache.spark.sql.catalyst.expressions.GroupingID groupingID, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupByExprs) static Throwablestatic ThrowablegroupingSizeTooLargeError(int sizeLimit) static Throwablestatic ThrowablehiveTableTypeUnsupportedError(String tableName, String tableType) static Throwablestatic ThrowableidentifierTooManyNamePartsError(String originalIdentifier) static ThrowableillegalParquetTypeError(String parquetType) static ThrowableincompatibleDataToTableAmbiguousColumnNameError(String tableName, String colName) static ThrowableincompatibleDataToTableCannotFindDataError(String tableName, String colName) static ThrowableincompatibleDataToTableCannotSafelyCastError(String tableName, String colName, String srcType, String targetType) static ThrowableincompatibleDataToTableExtraColumnsError(String tableName, String extraColumns) static ThrowableincompatibleDataToTableExtraStructFieldsError(String tableName, String colName, String extraFields) static ThrowableincompatibleDataToTableNullableArrayElementsError(String tableName, String colName) static ThrowableincompatibleDataToTableNullableColumnError(String tableName, String colName) static ThrowableincompatibleDataToTableNullableMapValuesError(String tableName, String colName) static ThrowableincompatibleDataToTableStructMissingFieldsError(String tableName, String colName, String missingFields) static ThrowableincompatibleDataToTableUnexpectedColumnNameError(String tableName, String colName, int order, String expected, String found) static ThrowableincompatibleViewSchemaChangeError(String viewName, String colName, int expectedNum, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> actualCols, scala.Option<String> viewDDL) static ThrowableinputSourceDiffersFromDataSourceProviderError(String source, String tableName, org.apache.spark.sql.catalyst.catalog.CatalogTable table) static ThrowableinsertIntoViewNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableinsertMismatchedColumnNumberError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> targetAttributes, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> sourceAttributes, int staticPartitionsSize) static ThrowableinsertMismatchedPartitionNumberError(StructType targetPartitionSchema, int providedPartitionsSize) static Throwablestatic ThrowableinsufficientTablePropertyPartError(String key, String totalAmountOfParts) static ThrowableinvalidBoundaryEndError(long end) static ThrowableinvalidBoundaryStartError(long start) static ThrowableinvalidBucketNumberError(int bucketingMaxBuckets, int numBuckets) static ThrowableinvalidBucketsNumberError(String numBuckets, String e) static ThrowableinvalidByteStringFormatError(Object unsupported) static ThrowableinvalidCoalesceHintParameterError(String hintName) static ThrowableinvalidColumnNameAsPathError(String datasource, String columnName) static ThrowableinvalidColumnOrFieldDataTypeError(scala.collection.Seq<String> name, DataType dt, DataType expected) static ThrowableinvalidDatabaseNameError(String quoted) static ThrowableinvalidDataSourceError(String className) static ThrowableinvalidEscapeChar(org.apache.spark.sql.catalyst.expressions.Expression sqlExpr) static ThrowableinvalidFieldName(scala.collection.Seq<String> fieldName, scala.collection.Seq<String> path, org.apache.spark.sql.catalyst.trees.Origin context) static Throwablestatic ThrowableinvalidFileFormatForStoredAsError(org.apache.spark.sql.catalyst.plans.logical.SerdeInfo serdeInfo) static ThrowableinvalidHintParameterError(String hintName, scala.collection.Seq<Object> invalidParams) static Throwablestatic ThrowableinvalidJoinTypeInJoinWithError(org.apache.spark.sql.catalyst.plans.JoinType joinType) static ThrowableinvalidJsonSchema(DataType schema) static Throwablestatic Throwablestatic ThrowableinvalidOrderingForConstantValuePartitionColumnError(StructType targetPartitionSchema) static ThrowableinvalidPandasUDFPlacementError(scala.collection.Seq<String> groupAggPandasUDFNames) static ThrowableinvalidPartitionColumnError(String partKey, StructType targetPartitionSchema) static ThrowableinvalidPartitionColumnKeyInTableError(String key, String tblName) static Throwablestatic ThrowableinvalidPartitionSpecError(String details) static ThrowableinvalidPartitionSpecError(String specKeys, scala.collection.Seq<String> partitionColumnNames, String tableName) static ThrowableinvalidPartitionTransformationError(org.apache.spark.sql.catalyst.expressions.Expression expr) static ThrowableinvalidRepartitionExpressionsError(scala.collection.Seq<Object> sortOrders) static ThrowableinvalidRowLevelOperationAssignments(scala.collection.Seq<org.apache.spark.sql.catalyst.plans.logical.Assignment> assignments, scala.collection.Seq<String> errors) static ThrowableinvalidStarUsageError(String prettyName, scala.collection.Seq<org.apache.spark.sql.catalyst.analysis.Star> stars) static ThrowableinvalidTimestampExprForTimeTravel(String errorClass, org.apache.spark.sql.catalyst.expressions.Expression expr) static ThrowableinvalidTimestampProvidedForStrategyError(String strategy, String timeString) static Throwablestatic ThrowableinvalidViewNameError(String viewName) static ThrowableinvalidViewText(String viewText, String viewName) static ThrowablejoinConditionMissingOrTrivialError(org.apache.spark.sql.catalyst.plans.logical.Join join, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan left, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan right) static ThrowablejoinStrategyHintParameterNotSupportedError(Object unsupported) static ThrowablekeyValueInMapNotStringError(org.apache.spark.sql.catalyst.expressions.CreateMap m) static ThrowablelateralColumnAliasInAggFuncUnsupportedError(scala.collection.Seq<String> lcaNameParts, org.apache.spark.sql.catalyst.expressions.Expression aggExpr) static ThrowablelateralColumnAliasInAggWithWindowAndHavingUnsupportedError(scala.collection.Seq<String> lcaNameParts) static ThrowablelateralColumnAliasInWindowUnsupportedError(scala.collection.Seq<String> lcaNameParts, org.apache.spark.sql.catalyst.expressions.Expression windowExpr) static Throwablestatic ThrowableliteralTypeUnsupportedForSourceTypeError(String field, org.apache.spark.sql.catalyst.expressions.Expression source) static Throwablestatic ThrowableloadDataNotSupportedForDatasourceTablesError(String tableIdentWithDB) static Throwablestatic ThrowableloadDataPartitionSizeNotMatchNumPartitionColumnsError(String tableIdentWithDB, int partitionSize, int targetTableSize) static ThrowableloadDataTargetTableNotPartitionedButPartitionSpecWasProvidedError(String tableIdentWithDB) static ThrowableloadDataWithoutPartitionSpecProvidedError(String tableIdentWithDB) static Throwablestatic ThrowablemismatchedTableBucketingError(String tableName, String specifiedBucketString, String existingBucketString) static ThrowablemismatchedTableColumnNumberError(String tableName, org.apache.spark.sql.catalyst.catalog.CatalogTable existingTable, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan query) static ThrowablemismatchedTableFormatError(String tableName, Class<?> existingProvider, Class<?> specifiedProvider) static ThrowablemismatchedTableLocationError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.catalog.CatalogTable existingTable, org.apache.spark.sql.catalyst.catalog.CatalogTable tableDesc) static ThrowablemismatchedTablePartitionColumnError(String tableName, scala.collection.Seq<String> specifiedPartCols, String existingPartCols) static ThrowablemissingCatalogAbilityError(CatalogPlugin plugin, String ability) static ThrowablemissingFieldError(scala.collection.Seq<String> fieldName, org.apache.spark.sql.catalyst.analysis.ResolvedTable table, org.apache.spark.sql.catalyst.trees.Origin context) static ThrowablemissingStaticPartitionColumn(String staticName) static ThrowablemixedRefsInAggFunc(String funcStr, org.apache.spark.sql.catalyst.trees.Origin origin) static ThrowablemoreThanOneGeneratorError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> generators, String clause) static ThrowablemultiplePartitionColumnValuesSpecifiedError(StructField field, scala.collection.immutable.Map<String, String> potentialSpecs) static ThrowablemultiTimeWindowExpressionsNotSupportedError(org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowablenamedArgumentsNotEnabledError(String functionName, String argumentName) static ThrowablenamedArgumentsNotSupported(String functionName) static ThrowablenamespaceAlreadyExistsError(String[] namespace) static Throwablestatic ThrowablenestedGeneratorError(org.apache.spark.sql.catalyst.expressions.Expression trimmedNestedGenerator) static ThrowablenoHandlerForUDAFError(String name) static Throwablestatic Throwablestatic ThrowablenonDeterministicMergeCondition(String condName, org.apache.spark.sql.catalyst.expressions.Expression cond) static ThrowablenonLiteralPivotValError(org.apache.spark.sql.catalyst.expressions.Expression pivotVal) static Throwablestatic ThrowablenonPartitionColError(String partitionName) static ThrowablenonPartitionPruningPredicatesNotExpectedError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> nonPartitionPruningPredicates) static ThrowablenoSuchFunctionError(org.apache.spark.sql.catalyst.FunctionIdentifier identifier) static ThrowablenoSuchNamespaceError(String[] namespace) static ThrowablenoSuchPartitionError(String db, String table, scala.collection.immutable.Map<String, String> partition) static ThrowablenoSuchStructFieldInGivenFieldsError(String fieldName, StructField[] fields) static ThrowablenoSuchTableError(String db, String table) static org.apache.spark.sql.catalyst.analysis.NoSuchTableExceptionnoSuchTableError(Identifier ident) static ThrowablenoSuchTableError(scala.collection.Seq<String> nameParts) static ThrowablenotAllowedToCreatePermanentViewByReferencingTempFuncError(org.apache.spark.sql.catalyst.TableIdentifier name, String funcName) static ThrowablenotAllowedToCreatePermanentViewByReferencingTempViewError(org.apache.spark.sql.catalyst.TableIdentifier name, String nameParts) static ThrowablenotAllowedToCreatePermanentViewWithoutAssigningAliasForExpressionError(org.apache.spark.sql.catalyst.TableIdentifier name, org.apache.spark.sql.catalyst.expressions.Attribute attr) static ThrowablenotNullConstraintViolationArrayElementError(scala.collection.Seq<String> path) static ThrowablenotNullConstraintViolationMapValueError(scala.collection.Seq<String> path) static ThrowablenullableColumnOrFieldError(scala.collection.Seq<String> name) static ThrowablenullableRowIdError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> nullableRowIdAttrs) static Throwablestatic Throwablestatic ThrowableoperationNotSupportPartitioningError(String operation) static ThrowableoptionMustBeConstant(String key, scala.Option<Throwable> cause) static Throwablestatic Throwablestatic ThrowableorderByPositionRangeError(int index, int size, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowableouterScopeFailureForNewInstanceError(String className) static ThrowableoutputPathAlreadyExistsError(org.apache.hadoop.fs.Path outputPath) static Throwablestatic ThrowableparquetTypeUnsupportedYetError(String parquetType) static ThrowableparseModeUnsupportedError(String funcName, org.apache.spark.sql.catalyst.util.ParseMode mode) static Throwablestatic ThrowablepartitionColumnNotFoundInSchemaError(String col, String schemaCatalog) static ThrowablepartitionColumnNotSpecifiedError(String format, String partitionColumn) static ThrowablepartitionNotSpecifyLocationUriError(String specString) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowablepivotValDataTypeMismatchError(org.apache.spark.sql.catalyst.expressions.Expression pivotVal, org.apache.spark.sql.catalyst.expressions.Expression pivotCol) static Throwablestatic ThrowablepositionalAndNamedArgumentDoubleReference(String functionName, String parameterName) static ThrowableprotobufClassLoadError(String protobufClassName, String explanation, Throwable cause) static ThrowableprotobufDescriptorDependencyError(String dependencyName) static ThrowableprotobufFieldMatchError(String field, String protobufSchema, String matchSize, String matches) static Throwablestatic ThrowableprotobufTypeUnsupportedYetError(String protobufType) static Throwablestatic Throwablestatic ThrowablereadNonStreamingTempViewError(String quoted) static ThrowablerecoverQueryFromCheckpointUnsupportedError(org.apache.hadoop.fs.Path checkpointPath) static ThrowablerecursiveViewDetectedError(org.apache.spark.sql.catalyst.TableIdentifier viewIdent, scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> newPath) static ThrowablereferenceColNotFoundForAlterTableChangesError(String fieldName, String[] fields) static Throwablestatic ThrowablerenameTempViewToExistingViewError(String newName) static Throwablestatic ThrowablerequestedPartitionsMismatchTablePartitionsError(String tableName, scala.collection.immutable.Map<String, scala.Option<String>> normalizedPartSpec, StructType partColNames) static ThrowablerequiredParameterNotFound(String functionName, String parameterName, int index) static ThrowablerequireLiteralParameter(String funcName, String argName, String requiredType) static ThrowablerequiresSinglePartNamespaceError(scala.collection.Seq<String> namespace) static AnalysisExceptionresolveException(String colName, String[] fields) static ThrowableresourceTypeNotSupportedError(String resourceType) static Throwablestatic ThrowableschemaIsNotStructTypeError(org.apache.spark.sql.catalyst.expressions.Expression exp, DataType dataType) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowableselectExprNotInGroupByError(org.apache.spark.sql.catalyst.expressions.Expression expr, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Alias> groupByAliases) static Throwablestatic Throwablestatic Throwablestatic ThrowableshowColumnsWithConflictDatabasesError(scala.collection.Seq<String> db, org.apache.spark.sql.catalyst.TableIdentifier v1TableName) static ThrowableshowCreateTableAsSerdeNotAllowedOnSparkDataSourceTableError(org.apache.spark.sql.catalyst.TableIdentifier table) static Throwablestatic ThrowableshowCreateTableFailToExecuteUnsupportedConfError(org.apache.spark.sql.catalyst.TableIdentifier table, scala.collection.mutable.StringBuilder builder) static ThrowableshowCreateTableFailToExecuteUnsupportedFeatureError(org.apache.spark.sql.catalyst.catalog.CatalogTable table) static Throwablestatic ThrowableshowCreateTableNotSupportTransactionalHiveTableError(org.apache.spark.sql.catalyst.catalog.CatalogTable table) static ThrowableshowCreateTableOrViewFailToExecuteUnsupportedFeatureError(org.apache.spark.sql.catalyst.catalog.CatalogTable table, scala.collection.Seq<String> features) static ThrowableshowPartitionNotAllowedOnTableNotPartitionedError(String tableIdentWithDB) static ThrowablesingleTableStarInCountNotAllowedError(String targetString) static Throwablestatic Throwablestatic Throwablestatic ThrowablestarExpandDataTypeNotSupportedError(scala.collection.Seq<String> attributes) static Throwablestatic ThrowablestaticPartitionInUserSpecifiedColumnsError(String staticName) static ThrowablestreamingIntoViewNotSupportedError(String viewName) static ThrowablestreamingSourcesDoNotSupportCommonExecutionModeError(scala.collection.Seq<String> microBatchSources, scala.collection.Seq<String> continuousSources) static ThrowablestreamJoinStreamWithoutEqualityPredicateUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) static ThrowablesubqueryNotAllowedInMergeCondition(String condName, org.apache.spark.sql.catalyst.expressions.Expression cond) static ThrowabletableAlreadyExistsError(String table) static ThrowabletableAlreadyExistsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) static Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic Throwablestatic ThrowabletableIdentifierExistsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdentifier) static ThrowabletableIdentifierNotConvertedToHadoopFsRelationError(org.apache.spark.sql.catalyst.TableIdentifier tableIdentifier) static ThrowabletableIndexNotSupportedError(String errorMessage) static Throwablestatic ThrowabletableIsNotViewError(org.apache.spark.sql.catalyst.TableIdentifier name) static ThrowabletableNotSpecifyDatabaseError(org.apache.spark.sql.catalyst.TableIdentifier identifier) static ThrowabletableNotSpecifyLocationUriError(org.apache.spark.sql.catalyst.TableIdentifier identifier) static ThrowabletableNotSupportStreamingWriteError(String tableName, Table t) static Throwablestatic ThrowabletableOrViewNotFound(scala.collection.Seq<String> ident) static Throwablestatic ThrowabletempViewNotCachedForAnalyzingColumnsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) static ThrowabletempViewNotSupportStreamingWriteError(String viewName) static Throwablestatic ThrowabletimeTravelUnsupportedError(String relationId) static StringtoDSOption(String option) static Stringstatic StringtoSQLConfVal(String conf) static StringtoSQLExpr(org.apache.spark.sql.catalyst.expressions.Expression e) static Stringstatic Stringstatic StringtoSQLSchema(String schema) static Stringstatic Stringstatic StringtoSQLType(org.apache.spark.sql.types.AbstractDataType t) static StringtoSQLValue(double value) static StringtoSQLValue(float value) static StringtoSQLValue(int value) static StringtoSQLValue(long value) static StringtoSQLValue(short value) static StringtoSQLValue(Object v, DataType t) static StringtoSQLValue(String value) static StringtoSQLValue(org.apache.spark.unsafe.types.UTF8String value) static ThrowabletruncateTableOnExternalTablesError(String tableIdentWithDB) static ThrowabletruncateTablePartitionNotSupportedForNotPartitionedTablesError(String tableIdentWithDB) static Throwablestatic ThrowableudfClassImplementMultiUDFInterfacesError(String className) static Throwablestatic ThrowableunableToLocateProtobufMessageError(String messageName) static ThrowableunexpectedInputDataTypeError(String functionName, int paramIndex, DataType dataType, org.apache.spark.sql.catalyst.expressions.Expression expression) static ThrowableunexpectedPartitionColumnPrefixError(String table, String database, String schemaColumns, String specColumns) static ThrowableunexpectedPositionalArgument(String functionName, String precedingNamedArgument) static ThrowableunexpectedRequiredParameterInFunctionSignature(String functionName, org.apache.spark.sql.catalyst.plans.logical.FunctionSignature functionSignature) static ThrowableunexpectedSchemaTypeError(org.apache.spark.sql.catalyst.expressions.Expression exp) static ThrowableunknownHiveResourceTypeError(String resourceType) static ThrowableunknownProtobufMessageTypeError(String descriptorName, String containingType) static ThrowableunorderablePivotColError(org.apache.spark.sql.catalyst.expressions.Expression pivotCol) static ThrowableunpivotRequiresAttributes(String given, String empty, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> expressions) static Throwablestatic ThrowableunpivotValueDataTypeMismatchError(scala.collection.Seq<scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression>> values) static ThrowableunpivotValueSizeMismatchError(int names) static ThrowableunrecognizedParameterName(String functionName, String argumentName, scala.collection.Seq<String> candidates) static Throwablestatic ThrowableunresolvedAttributeError(String errorClass, String colName, scala.collection.Seq<String> candidates, org.apache.spark.sql.catalyst.trees.Origin origin) static ThrowableunresolvedColumnError(String columnName, scala.collection.Seq<String> proposal) static AnalysisExceptionunresolvedColumnWithSuggestionError(String objectName, String suggestion) static ThrowableunresolvedFieldError(String fieldName, scala.collection.Seq<String> columnPath, scala.collection.Seq<String> proposal) static ThrowableunresolvedRoutineError(org.apache.spark.sql.catalyst.FunctionIdentifier name, scala.collection.Seq<String> searchPath) static ThrowableunresolvedRoutineError(scala.collection.Seq<String> nameParts, scala.collection.Seq<String> searchPath, org.apache.spark.sql.catalyst.trees.Origin context) static ThrowableunresolvedUsingColForJoinError(String colName, String suggestion, String side) static ThrowableunsetNonExistentPropertiesError(scala.collection.Seq<String> properties, org.apache.spark.sql.catalyst.TableIdentifier table) static Throwablestatic ThrowableunsupportedBatchReadError(Table table) static ThrowableunsupportedCorrelatedReferenceDataTypeError(org.apache.spark.sql.catalyst.expressions.Expression expr, DataType dataType, org.apache.spark.sql.catalyst.trees.Origin origin) static ThrowableunsupportedDeleteByConditionWithSubqueryError(org.apache.spark.sql.catalyst.expressions.Expression condition) static Throwablestatic ThrowableunsupportedIfNotExistsError(String tableName) static ThrowableunsupportedJDBCNamespaceChangeInCatalogError(scala.collection.Seq<NamespaceChange> changes) static Throwablestatic Throwablestatic Throwablestatic ThrowableunsupportedTableOperationError(org.apache.spark.sql.catalyst.TableIdentifier ident, String operation) static ThrowableunsupportedTableOperationError(CatalogPlugin catalog, Identifier ident, String operation) static Throwablestatic ThrowableupCastFailureError(String fromStr, org.apache.spark.sql.catalyst.expressions.Expression from, DataType to, scala.collection.Seq<String> walkedTypePath) static ThrowableusePythonUDFInJoinConditionUnsupportedError(org.apache.spark.sql.catalyst.plans.JoinType joinType) static ThrowableuserDefinedPartitionNotFoundInJDBCRelationError(String columnName, String schema) static ThrowableuserSpecifiedSchemaMismatchActualSchemaError(StructType schema, StructType actualSchema) static ThrowableuserSpecifiedSchemaUnsupportedError(String operation) static Throwablestatic Throwablev2FunctionInvalidInputTypeLengthError(BoundFunction bound, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> args) static ThrowableviewAlreadyExistsError(org.apache.spark.sql.catalyst.TableIdentifier name) static ThrowableviewDepthExceedsMaxResolutionDepthError(org.apache.spark.sql.catalyst.TableIdentifier identifier, int maxNestedViewDepth, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static Throwablestatic ThrowablewindowFrameNotMatchRequiredFrameError(org.apache.spark.sql.catalyst.expressions.SpecifiedWindowFrame f, org.apache.spark.sql.catalyst.expressions.WindowFrame required) static Throwablestatic Throwablestatic ThrowablewindowFunctionNotAllowedError(String clauseName) static ThrowablewindowFunctionWithWindowFrameNotOrderedError(org.apache.spark.sql.catalyst.expressions.WindowFunction wf) static ThrowablewindowSpecificationNotDefinedError(String windowName) static Throwablestatic Throwablestatic Throwablestatic ThrowablewriteIntoV1TableNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowablewriteIntoViewNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) static ThrowablewriteWithSaveModeUnsupportedBySourceError(String source, String createMode) static ThrowablewrongCommandForObjectTypeError(String operation, String requiredType, String objectName, String foundType, String alternative) static ThrowablewrongNumArgsError(String name, scala.collection.Seq<Object> validParametersCount, int actualNumber, String legacyNum, String legacyConfKey, String legacyConfValue) static Throwable
-
Constructor Details
-
QueryCompilationErrors
public QueryCompilationErrors()
-
-
Method Details
-
unexpectedRequiredParameterInFunctionSignature
-
namedArgumentsNotSupported
-
positionalAndNamedArgumentDoubleReference
-
doubleNamedArgumentReference
-
requiredParameterNotFound
-
unrecognizedParameterName
-
unexpectedPositionalArgument
-
groupingIDMismatchError
public static Throwable groupingIDMismatchError(org.apache.spark.sql.catalyst.expressions.GroupingID groupingID, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupByExprs) -
groupingColInvalidError
public static Throwable groupingColInvalidError(org.apache.spark.sql.catalyst.expressions.Expression groupingCol, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> groupByExprs) -
groupingSizeTooLargeError
-
zeroArgumentIndexError
-
unorderablePivotColError
public static Throwable unorderablePivotColError(org.apache.spark.sql.catalyst.expressions.Expression pivotCol) -
nonLiteralPivotValError
public static Throwable nonLiteralPivotValError(org.apache.spark.sql.catalyst.expressions.Expression pivotVal) -
pivotValDataTypeMismatchError
public static Throwable pivotValDataTypeMismatchError(org.apache.spark.sql.catalyst.expressions.Expression pivotVal, org.apache.spark.sql.catalyst.expressions.Expression pivotCol) -
unpivotRequiresAttributes
-
unpivotRequiresValueColumns
-
unpivotValueSizeMismatchError
-
unpivotValueDataTypeMismatchError
public static Throwable unpivotValueDataTypeMismatchError(scala.collection.Seq<scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression>> values) -
unsupportedIfNotExistsError
-
nonPartitionColError
-
missingStaticPartitionColumn
-
staticPartitionInUserSpecifiedColumnsError
-
nestedGeneratorError
public static Throwable nestedGeneratorError(org.apache.spark.sql.catalyst.expressions.Expression trimmedNestedGenerator) -
moreThanOneGeneratorError
-
generatorOutsideSelectError
public static Throwable generatorOutsideSelectError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) -
legacyStoreAssignmentPolicyError
-
namedArgumentsNotEnabledError
-
unresolvedUsingColForJoinError
-
unresolvedAttributeError
-
unresolvedColumnError
-
unresolvedFieldError
-
dataTypeMismatchForDeserializerError
-
fieldNumberMismatchForDeserializerError
-
upCastFailureError
-
outerScopeFailureForNewInstanceError
-
referenceColNotFoundForAlterTableChangesError
-
windowSpecificationNotDefinedError
-
selectExprNotInGroupByError
public static Throwable selectExprNotInGroupByError(org.apache.spark.sql.catalyst.expressions.Expression expr, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Alias> groupByAliases) -
groupingMustWithGroupingSetsOrCubeOrRollupError
-
pandasUDFAggregateNotSupportedInPivotError
-
aggregateExpressionRequiredForPivotError
-
writeIntoTempViewNotAllowedError
-
readNonStreamingTempViewError
-
viewDepthExceedsMaxResolutionDepthError
public static Throwable viewDepthExceedsMaxResolutionDepthError(org.apache.spark.sql.catalyst.TableIdentifier identifier, int maxNestedViewDepth, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
insertIntoViewNotAllowedError
public static Throwable insertIntoViewNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
writeIntoViewNotAllowedError
public static Throwable writeIntoViewNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
writeIntoV1TableNotAllowedError
public static Throwable writeIntoV1TableNotAllowedError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
expectTableNotViewError
-
expectViewNotTempViewError
-
expectViewNotTableError
-
expectTableOrPermanentViewNotTempViewError
-
expectPersistentFuncError
-
permanentViewNotSupportedByStreamingReadingAPIError
-
starNotAllowedWhenGroupByOrdinalPositionUsedError
-
invalidStarUsageError
-
singleTableStarInCountNotAllowedError
-
orderByPositionRangeError
public static Throwable orderByPositionRangeError(int index, int size, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
groupByPositionRefersToAggregateFunctionError
public static Throwable groupByPositionRefersToAggregateFunctionError(int index, org.apache.spark.sql.catalyst.expressions.Expression expr) -
groupByPositionRangeError
-
generatorNotExpectedError
-
functionWithUnsupportedSyntaxError
-
nonDeterministicFilterInAggregateError
-
nonBooleanFilterInAggregateError
-
aggregateInAggregateFilterError
-
windowFunctionInAggregateFilterError
-
aliasNumberNotMatchColumnNumberError
public static Throwable aliasNumberNotMatchColumnNumberError(int columnSize, int outputSize, org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
aliasesNumberNotMatchUDTFOutputError
-
windowAggregateFunctionWithFilterNotSupportedError
-
windowFunctionInsideAggregateFunctionNotAllowedError
-
expressionWithoutWindowExpressionError
public static Throwable expressionWithoutWindowExpressionError(org.apache.spark.sql.catalyst.expressions.NamedExpression expr) -
expressionWithMultiWindowExpressionsError
public static Throwable expressionWithMultiWindowExpressionsError(org.apache.spark.sql.catalyst.expressions.NamedExpression expr, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.WindowSpecDefinition> distinctWindowSpec) -
windowFunctionNotAllowedError
-
cannotSpecifyWindowFrameError
-
windowFrameNotMatchRequiredFrameError
public static Throwable windowFrameNotMatchRequiredFrameError(org.apache.spark.sql.catalyst.expressions.SpecifiedWindowFrame f, org.apache.spark.sql.catalyst.expressions.WindowFrame required) -
windowFunctionWithWindowFrameNotOrderedError
public static Throwable windowFunctionWithWindowFrameNotOrderedError(org.apache.spark.sql.catalyst.expressions.WindowFunction wf) -
multiTimeWindowExpressionsNotSupportedError
public static Throwable multiTimeWindowExpressionsNotSupportedError(org.apache.spark.sql.catalyst.trees.TreeNode<?> t) -
sessionWindowGapDurationDataTypeError
-
unresolvedRoutineError
-
unresolvedRoutineError
-
wrongNumArgsError
-
alterV2TableSetLocationWithPartitionNotSupportedError
-
joinStrategyHintParameterNotSupportedError
-
invalidHintParameterError
-
invalidCoalesceHintParameterError
-
attributeNameSyntaxError
-
starExpandDataTypeNotSupportedError
-
cannotResolveStarExpandGivenInputColumnsError
-
addColumnWithV1TableCannotSpecifyNotNullError
-
unsupportedTableOperationError
public static Throwable unsupportedTableOperationError(CatalogPlugin catalog, Identifier ident, String operation) -
unsupportedTableOperationError
-
catalogOperationNotSupported
-
alterColumnWithV1TableCannotSpecifyNotNullError
-
alterColumnCannotFindColumnInV1TableError
-
invalidDatabaseNameError
-
wrongCommandForObjectTypeError
-
showColumnsWithConflictDatabasesError
-
cannotCreateTableWithBothProviderAndSerdeError
-
invalidFileFormatForStoredAsError
public static Throwable invalidFileFormatForStoredAsError(org.apache.spark.sql.catalyst.plans.logical.SerdeInfo serdeInfo) -
commandNotSupportNestedColumnError
-
renameTempViewToExistingViewError
-
cannotDropNonemptyDatabaseError
-
cannotDropNonemptyNamespaceError
-
invalidNameForTableOrDatabaseError
-
cannotCreateDatabaseWithSameNameAsPreservedDatabaseError
-
cannotDropDefaultDatabaseError
-
cannotUsePreservedDatabaseAsCurrentDatabaseError
-
createExternalTableWithoutLocationError
-
dropNonExistentColumnsNotSupportedError
-
cannotRetrieveTableOrViewNotInSameDatabaseError
public static Throwable cannotRetrieveTableOrViewNotInSameDatabaseError(scala.collection.Seq<org.apache.spark.sql.catalyst.QualifiedTableName> qualifiedTableNames) -
renameTableSourceAndDestinationMismatchError
-
cannotRenameTempViewWithDatabaseSpecifiedError
public static Throwable cannotRenameTempViewWithDatabaseSpecifiedError(org.apache.spark.sql.catalyst.TableIdentifier oldName, org.apache.spark.sql.catalyst.TableIdentifier newName) -
cannotRenameTempViewToExistingTableError
public static Throwable cannotRenameTempViewToExistingTableError(org.apache.spark.sql.catalyst.TableIdentifier newName) -
invalidPartitionSpecError
-
functionAlreadyExistsError
public static Throwable functionAlreadyExistsError(org.apache.spark.sql.catalyst.FunctionIdentifier func) -
cannotLoadClassWhenRegisteringFunctionError
-
resourceTypeNotSupportedError
-
tableNotSpecifyDatabaseError
public static Throwable tableNotSpecifyDatabaseError(org.apache.spark.sql.catalyst.TableIdentifier identifier) -
tableNotSpecifyLocationUriError
public static Throwable tableNotSpecifyLocationUriError(org.apache.spark.sql.catalyst.TableIdentifier identifier) -
partitionNotSpecifyLocationUriError
-
invalidBucketNumberError
-
corruptedTableNameContextInCatalogError
-
corruptedViewSQLConfigsInCatalogError
-
corruptedViewQueryOutputColumnsInCatalogError
-
corruptedViewReferredTempViewInCatalogError
-
corruptedViewReferredTempFunctionsInCatalogError
-
columnStatisticsDeserializationNotSupportedError
-
columnStatisticsSerializationNotSupportedError
-
insufficientTablePropertyError
-
insufficientTablePropertyPartError
-
unexpectedSchemaTypeError
public static Throwable unexpectedSchemaTypeError(org.apache.spark.sql.catalyst.expressions.Expression exp) -
schemaIsNotStructTypeError
-
keyValueInMapNotStringError
public static Throwable keyValueInMapNotStringError(org.apache.spark.sql.catalyst.expressions.CreateMap m) -
nonMapFunctionNotAllowedError
-
invalidFieldTypeForCorruptRecordError
-
dataTypeUnsupportedByClassError
-
parseModeUnsupportedError
-
requireLiteralParameter
-
literalTypeUnsupportedForSourceTypeError
-
arrayComponentTypeUnsupportedError
-
secondArgumentNotDoubleLiteralError
-
dataTypeUnsupportedByExtractValueError
-
noHandlerForUDAFError
-
batchWriteCapabilityError
-
unsupportedDeleteByConditionWithSubqueryError
public static Throwable unsupportedDeleteByConditionWithSubqueryError(org.apache.spark.sql.catalyst.expressions.Expression condition) -
cannotTranslateExpressionToSourceFilterError
public static Throwable cannotTranslateExpressionToSourceFilterError(org.apache.spark.sql.catalyst.expressions.Expression f) -
cannotDeleteTableWhereFiltersError
-
describeDoesNotSupportPartitionForV2TablesError
-
cannotReplaceMissingTableError
-
cannotReplaceMissingTableError
public static Throwable cannotReplaceMissingTableError(Identifier tableIdentifier, scala.Option<Throwable> cause) -
unsupportedBatchReadError
-
unsupportedMicroBatchOrContinuousScanError
-
unsupportedAppendInBatchModeError
-
unsupportedDynamicOverwriteInBatchModeError
-
unsupportedTruncateInBatchModeError
-
unsupportedOverwriteByFilterInBatchModeError
-
streamingSourcesDoNotSupportCommonExecutionModeError
-
noSuchTableError
public static org.apache.spark.sql.catalyst.analysis.NoSuchTableException noSuchTableError(Identifier ident) -
noSuchTableError
-
noSuchNamespaceError
-
tableAlreadyExistsError
-
requiresSinglePartNamespaceError
-
namespaceAlreadyExistsError
-
cannotCreateJDBCTableUsingProviderError
-
cannotCreateJDBCTableUsingLocationError
-
cannotCreateJDBCNamespaceUsingProviderError
-
cannotCreateJDBCNamespaceWithPropertyError
-
cannotSetJDBCNamespaceWithPropertyError
-
cannotUnsetJDBCNamespaceWithPropertyError
-
unsupportedJDBCNamespaceChangeInCatalogError
public static Throwable unsupportedJDBCNamespaceChangeInCatalogError(scala.collection.Seq<NamespaceChange> changes) -
tableDoesNotSupportReadsError
-
tableDoesNotSupportWritesError
-
tableDoesNotSupportDeletesError
-
tableDoesNotSupportTruncatesError
-
tableDoesNotSupportPartitionManagementError
-
tableDoesNotSupportAtomicPartitionManagementError
-
tableIsNotRowLevelOperationTableError
-
cannotRenameTableWithAlterViewError
-
analyzeTableNotSupportedForV2TablesError
-
alterTableRecoverPartitionsNotSupportedForV2TablesError
-
alterTableSerDePropertiesNotSupportedForV2TablesError
-
loadDataNotSupportedForV2TablesError
-
showCreateTableAsSerdeNotSupportedForV2TablesError
-
showColumnsNotSupportedForV2TablesError
-
repairTableNotSupportedForV2TablesError
-
databaseFromV1SessionCatalogNotSpecifiedError
-
nestedDatabaseUnsupportedByV1SessionCatalogError
-
invalidRepartitionExpressionsError
-
partitionColumnNotSpecifiedError
-
dataSchemaNotSpecifiedError
-
dataPathNotExistError
-
dataSourceOutputModeUnsupportedError
public static Throwable dataSourceOutputModeUnsupportedError(String className, OutputMode outputMode) -
schemaNotSpecifiedForSchemaRelationProviderError
-
userSpecifiedSchemaMismatchActualSchemaError
public static Throwable userSpecifiedSchemaMismatchActualSchemaError(StructType schema, StructType actualSchema) -
dataSchemaNotSpecifiedError
-
invalidDataSourceError
-
cannotSaveIntervalIntoExternalStorageError
-
cannotResolveAttributeError
-
orcNotUsedWithHiveEnabledError
-
failedToFindAvroDataSourceError
-
failedToFindKafkaDataSourceError
-
findMultipleDataSourceError
-
writeEmptySchemasUnsupportedByDataSourceError
-
insertMismatchedColumnNumberError
public static Throwable insertMismatchedColumnNumberError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> targetAttributes, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> sourceAttributes, int staticPartitionsSize) -
insertMismatchedPartitionNumberError
public static Throwable insertMismatchedPartitionNumberError(StructType targetPartitionSchema, int providedPartitionsSize) -
invalidPartitionColumnError
public static Throwable invalidPartitionColumnError(String partKey, StructType targetPartitionSchema) -
multiplePartitionColumnValuesSpecifiedError
public static Throwable multiplePartitionColumnValuesSpecifiedError(StructField field, scala.collection.immutable.Map<String, String> potentialSpecs) -
invalidOrderingForConstantValuePartitionColumnError
public static Throwable invalidOrderingForConstantValuePartitionColumnError(StructType targetPartitionSchema) -
cannotWriteDataToRelationsWithMultiplePathsError
-
failedToRebuildExpressionError
-
dataTypeUnsupportedByDataSourceError
-
failToResolveDataSourceForTableError
-
outputPathAlreadyExistsError
-
cannotUseDataTypeForPartitionColumnError
-
cannotUseAllColumnsForPartitionColumnsError
-
partitionColumnNotFoundInSchemaError
-
columnNotFoundInSchemaError
public static Throwable columnNotFoundInSchemaError(StructField col, scala.Option<StructType> tableSchema) -
saveDataIntoViewNotAllowedError
-
mismatchedTableFormatError
-
mismatchedTableLocationError
public static Throwable mismatchedTableLocationError(org.apache.spark.sql.catalyst.TableIdentifier identifier, org.apache.spark.sql.catalyst.catalog.CatalogTable existingTable, org.apache.spark.sql.catalyst.catalog.CatalogTable tableDesc) -
mismatchedTableColumnNumberError
-
cannotResolveColumnGivenInputColumnsError
-
mismatchedTablePartitionColumnError
-
mismatchedTableBucketingError
-
specifyPartitionNotAllowedWhenTableSchemaNotDefinedError
-
bucketingColumnCannotBePartOfPartitionColumnsError
-
bucketSortingColumnCannotBePartOfPartitionColumnsError
-
requestedPartitionsMismatchTablePartitionsError
public static Throwable requestedPartitionsMismatchTablePartitionsError(String tableName, scala.collection.immutable.Map<String, scala.Option<String>> normalizedPartSpec, StructType partColNames) -
ddlWithoutHiveSupportEnabledError
-
createTableColumnTypesOptionColumnNotFoundInSchemaError
public static Throwable createTableColumnTypesOptionColumnNotFoundInSchemaError(String col, StructType schema) -
parquetTypeUnsupportedYetError
-
illegalParquetTypeError
-
unrecognizedParquetTypeError
-
cannotConvertDataTypeToParquetTypeError
-
incompatibleViewSchemaChangeError
-
numberOfPartitionsNotAllowedWithUnspecifiedDistributionError
-
partitionSizeNotAllowedWithUnspecifiedDistributionError
-
numberAndSizeOfPartitionsNotAllowedTogether
-
unexpectedInputDataTypeError
-
streamJoinStreamWithoutEqualityPredicateUnsupportedError
public static Throwable streamJoinStreamWithoutEqualityPredicateUnsupportedError(org.apache.spark.sql.catalyst.plans.logical.LogicalPlan plan) -
invalidPandasUDFPlacementError
-
ambiguousAttributesInSelfJoinError
public static Throwable ambiguousAttributesInSelfJoinError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> ambiguousAttrs) -
ambiguousColumnOrFieldError
-
ambiguousReferenceError
-
cannotUseIntervalTypeInTableSchemaError
-
missingCatalogAbilityError
-
tableValuedFunctionTooManyTableArgumentsError
-
identifierTooManyNamePartsError
-
emptyMultipartIdentifierError
-
cannotOperateOnHiveDataSourceFilesError
-
setPathOptionAndCallWithPathParameterError
-
userSpecifiedSchemaUnsupportedError
-
tempViewNotSupportStreamingWriteError
-
streamingIntoViewNotSupportedError
-
inputSourceDiffersFromDataSourceProviderError
-
tableNotSupportStreamingWriteError
-
queryNameNotSpecifiedForMemorySinkError
-
sourceNotSupportedWithContinuousTriggerError
-
columnNotFoundInExistingColumnsError
-
operationNotSupportPartitioningError
-
mixedRefsInAggFunc
-
functionCannotProcessInputError
public static Throwable functionCannotProcessInputError(UnboundFunction unbound, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> arguments, UnsupportedOperationException unsupported) -
v2FunctionInvalidInputTypeLengthError
public static Throwable v2FunctionInvalidInputTypeLengthError(BoundFunction bound, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> args) -
commandUnsupportedInV2TableError
-
cannotResolveColumnNameAmongAttributesError
-
cannotWriteTooManyColumnsToTableError
-
cannotWriteNotEnoughColumnsToTableError
-
incompatibleDataToTableCannotFindDataError
-
incompatibleDataToTableAmbiguousColumnNameError
-
incompatibleDataToTableExtraColumnsError
-
incompatibleDataToTableExtraStructFieldsError
-
incompatibleDataToTableNullableColumnError
-
incompatibleDataToTableNullableArrayElementsError
-
incompatibleDataToTableNullableMapValuesError
-
incompatibleDataToTableCannotSafelyCastError
-
incompatibleDataToTableStructMissingFieldsError
-
incompatibleDataToTableUnexpectedColumnNameError
-
invalidRowLevelOperationAssignments
-
invalidEscapeChar
public static Throwable invalidEscapeChar(org.apache.spark.sql.catalyst.expressions.Expression sqlExpr) -
secondArgumentOfFunctionIsNotIntegerError
public static Throwable secondArgumentOfFunctionIsNotIntegerError(String function, NumberFormatException e) -
nonPartitionPruningPredicatesNotExpectedError
public static Throwable nonPartitionPruningPredicatesNotExpectedError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> nonPartitionPruningPredicates) -
columnNotDefinedInTableError
-
invalidLiteralForWindowDurationError
-
noSuchStructFieldInGivenFieldsError
-
ambiguousReferenceToFieldsError
-
secondArgumentInFunctionIsNotBooleanLiteralError
-
joinConditionMissingOrTrivialError
public static Throwable joinConditionMissingOrTrivialError(org.apache.spark.sql.catalyst.plans.logical.Join join, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan left, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan right) -
usePythonUDFInJoinConditionUnsupportedError
public static Throwable usePythonUDFInJoinConditionUnsupportedError(org.apache.spark.sql.catalyst.plans.JoinType joinType) -
conflictingAttributesInJoinConditionError
public static Throwable conflictingAttributesInJoinConditionError(org.apache.spark.sql.catalyst.expressions.AttributeSet conflictingAttrs, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan outerPlan, org.apache.spark.sql.catalyst.plans.logical.LogicalPlan subplan) -
emptyWindowExpressionError
public static Throwable emptyWindowExpressionError(org.apache.spark.sql.catalyst.plans.logical.Window expr) -
foundDifferentWindowFunctionTypeError
public static Throwable foundDifferentWindowFunctionTypeError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.NamedExpression> windowExpressions) -
charOrVarcharTypeAsStringUnsupportedError
-
escapeCharacterInTheMiddleError
-
escapeCharacterAtTheEndError
-
tableIdentifierExistsError
public static Throwable tableIdentifierExistsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdentifier) -
tableIdentifierNotConvertedToHadoopFsRelationError
public static Throwable tableIdentifierNotConvertedToHadoopFsRelationError(org.apache.spark.sql.catalyst.TableIdentifier tableIdentifier) -
alterDatabaseLocationUnsupportedError
-
hiveTableTypeUnsupportedError
-
hiveCreatePermanentFunctionsUnsupportedError
-
unknownHiveResourceTypeError
-
configRemovedInVersionError
-
invalidPartitionColumnKeyInTableError
-
invalidPartitionSpecError
-
columnAlreadyExistsError
-
noSuchTableError
-
tempViewNotCachedForAnalyzingColumnsError
public static Throwable tempViewNotCachedForAnalyzingColumnsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) -
columnTypeNotSupportStatisticsCollectionError
-
analyzeTableNotSupportedOnViewsError
-
unexpectedPartitionColumnPrefixError
-
noSuchPartitionError
-
analyzingColumnStatisticsNotSupportedForColumnTypeError
-
tableAlreadyExistsError
-
createTableAsSelectWithNonEmptyDirectoryError
-
unsetNonExistentPropertiesError
-
alterTableChangeColumnNotSupportedForColumnTypeError
public static Throwable alterTableChangeColumnNotSupportedForColumnTypeError(String tableName, StructField originColumn, StructField newColumn) -
cannotFindColumnError
-
alterTableSetSerdeForSpecificPartitionNotSupportedError
-
alterTableSetSerdeNotSupportedError
-
cmdOnlyWorksOnPartitionedTablesError
-
cmdOnlyWorksOnTableWithLocationError
-
actionNotAllowedOnTableWithFilesourcePartitionManagementDisabledError
-
actionNotAllowedOnTableSincePartitionMetadataNotStoredError
-
cannotAlterViewWithAlterTableError
-
cannotAlterTableWithAlterViewError
-
cannotOverwritePathBeingReadFromError
-
cannotOverwriteTableThatIsBeingReadFromError
public static Throwable cannotOverwriteTableThatIsBeingReadFromError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) -
cannotDropBuiltinFuncError
-
cannotRefreshBuiltInFuncError
-
cannotRefreshTempFuncError
-
noSuchFunctionError
public static Throwable noSuchFunctionError(org.apache.spark.sql.catalyst.FunctionIdentifier identifier) -
alterAddColNotSupportViewError
public static Throwable alterAddColNotSupportViewError(org.apache.spark.sql.catalyst.TableIdentifier table) -
alterAddColNotSupportDatasourceTableError
-
loadDataNotSupportedForDatasourceTablesError
-
loadDataWithoutPartitionSpecProvidedError
-
loadDataPartitionSizeNotMatchNumPartitionColumnsError
-
loadDataTargetTableNotPartitionedButPartitionSpecWasProvidedError
-
loadDataInputPathNotExistError
-
truncateTableOnExternalTablesError
-
truncateTablePartitionNotSupportedForNotPartitionedTablesError
-
failToTruncateTableWhenRemovingDataError
-
descPartitionNotAllowedOnTempView
-
descPartitionNotAllowedOnView
-
showPartitionNotAllowedOnTableNotPartitionedError
-
showCreateTableNotSupportedOnTempView
-
showCreateTableFailToExecuteUnsupportedFeatureError
public static Throwable showCreateTableFailToExecuteUnsupportedFeatureError(org.apache.spark.sql.catalyst.catalog.CatalogTable table) -
showCreateTableNotSupportTransactionalHiveTableError
public static Throwable showCreateTableNotSupportTransactionalHiveTableError(org.apache.spark.sql.catalyst.catalog.CatalogTable table) -
showCreateTableFailToExecuteUnsupportedConfError
public static Throwable showCreateTableFailToExecuteUnsupportedConfError(org.apache.spark.sql.catalyst.TableIdentifier table, scala.collection.mutable.StringBuilder builder) -
showCreateTableAsSerdeNotAllowedOnSparkDataSourceTableError
public static Throwable showCreateTableAsSerdeNotAllowedOnSparkDataSourceTableError(org.apache.spark.sql.catalyst.TableIdentifier table) -
showCreateTableOrViewFailToExecuteUnsupportedFeatureError
-
logicalPlanForViewNotAnalyzedError
-
cannotCreateViewTooManyColumnsError
-
cannotCreateViewNotEnoughColumnsError
-
tableIsNotViewError
-
viewAlreadyExistsError
-
createPersistedViewFromDatasetAPINotAllowedError
-
recursiveViewDetectedError
public static Throwable recursiveViewDetectedError(org.apache.spark.sql.catalyst.TableIdentifier viewIdent, scala.collection.Seq<org.apache.spark.sql.catalyst.TableIdentifier> newPath) -
notAllowedToCreatePermanentViewWithoutAssigningAliasForExpressionError
public static Throwable notAllowedToCreatePermanentViewWithoutAssigningAliasForExpressionError(org.apache.spark.sql.catalyst.TableIdentifier name, org.apache.spark.sql.catalyst.expressions.Attribute attr) -
notAllowedToCreatePermanentViewByReferencingTempViewError
-
notAllowedToCreatePermanentViewByReferencingTempFuncError
-
queryFromRawFilesIncludeCorruptRecordColumnError
-
userDefinedPartitionNotFoundInJDBCRelationError
-
invalidPartitionColumnTypeError
-
tableOrViewAlreadyExistsError
-
invalidColumnNameAsPathError
-
textDataSourceWithMultiColumnsError
-
cannotFindPartitionColumnInPartitionSchemaError
public static Throwable cannotFindPartitionColumnInPartitionSchemaError(StructField readField, StructType partitionSchema) -
cannotSpecifyDatabaseForTempViewError
public static Throwable cannotSpecifyDatabaseForTempViewError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) -
cannotCreateTempViewUsingHiveDataSourceError
-
invalidTimestampProvidedForStrategyError
-
hostOptionNotSetError
-
portOptionNotSetError
-
invalidIncludeTimestampValueError
-
checkpointLocationNotSpecifiedError
-
recoverQueryFromCheckpointUnsupportedError
public static Throwable recoverQueryFromCheckpointUnsupportedError(org.apache.hadoop.fs.Path checkpointPath) -
cannotFindColumnInRelationOutputError
-
invalidBoundaryStartError
-
invalidBoundaryEndError
-
tableOrViewNotFound
-
unsupportedTableChangeInJDBCCatalogError
-
pathOptionNotSetCorrectlyWhenReadingError
-
pathOptionNotSetCorrectlyWhenWritingError
-
writeWithSaveModeUnsupportedBySourceError
-
partitionByDoesNotAllowedWhenUsingInsertIntoError
-
cannotFindCatalogToHandleIdentifierError
-
sortByWithoutBucketingError
-
bucketByUnsupportedByOperationError
-
bucketByAndSortByUnsupportedByOperationError
-
tableAlreadyExistsError
public static Throwable tableAlreadyExistsError(org.apache.spark.sql.catalyst.TableIdentifier tableIdent) -
invalidPartitionTransformationError
public static Throwable invalidPartitionTransformationError(org.apache.spark.sql.catalyst.expressions.Expression expr) -
unresolvedColumnWithSuggestionError
public static AnalysisException unresolvedColumnWithSuggestionError(String objectName, String suggestion) -
resolveException
-
cannotParseIntervalError
-
invalidJoinTypeInJoinWithError
public static Throwable invalidJoinTypeInJoinWithError(org.apache.spark.sql.catalyst.plans.JoinType joinType) -
cannotPassTypedColumnInUntypedSelectError
-
invalidViewNameError
-
invalidBucketsNumberError
-
usingUntypedScalaUDFError
-
aggregationFunctionAppliedOnNonNumericColumnError
-
aggregationFunctionAppliedOnNonNumericColumnError
-
cannotModifyValueOfStaticConfigError
-
cannotModifyValueOfSparkConfigError
-
commandExecutionInRunnerUnsupportedError
-
udfClassDoesNotImplementAnyUDFInterfaceError
-
udfClassImplementMultiUDFInterfacesError
-
udfClassWithTooManyTypeArgumentsError
-
classWithoutPublicNonArgumentConstructorError
-
cannotLoadClassNotOnClassPathError
-
classDoesNotImplementUserDefinedAggregateFunctionError
-
missingFieldError
-
invalidFieldName
-
invalidJsonSchema
-
tableIndexNotSupportedError
-
invalidViewText
-
invalidTimeTravelSpecError
-
invalidTimestampExprForTimeTravel
-
timeTravelUnsupportedError
-
writeDistributionAndOrderingNotSupportedInContinuousExecution
-
defaultReferencesNotAllowedInComplexExpressionsInInsertValuesList
-
defaultReferencesNotAllowedInComplexExpressionsInUpdateSetClause
-
defaultReferencesNotAllowedInComplexExpressionsInMergeInsertsOrUpdates
-
nonDeterministicMergeCondition
-
subqueryNotAllowedInMergeCondition
-
aggregationNotAllowedInMergeCondition
-
failedToParseExistenceDefaultAsLiteral
-
defaultReferencesNotAllowedInDataSource
-
addNewDefaultColumnToExistingTableNotAllowed
-
defaultValuesDataTypeError
-
defaultValuesUnresolvedExprError
-
defaultValuesMayNotContainSubQueryExpressions
-
nullableColumnOrFieldError
-
notNullConstraintViolationArrayElementError
-
notNullConstraintViolationMapValueError
-
invalidColumnOrFieldDataTypeError
-
columnNotInGroupByClauseError
public static Throwable columnNotInGroupByClauseError(org.apache.spark.sql.catalyst.expressions.Expression expression) -
cannotConvertProtobufTypeToSqlTypeError
-
cannotConvertCatalystTypeToProtobufTypeError
-
cannotConvertProtobufTypeToCatalystTypeError
-
cannotConvertSqlTypeToProtobufError
-
protobufTypeUnsupportedYetError
-
unknownProtobufMessageTypeError
-
cannotFindCatalystTypeInProtobufSchemaError
-
cannotFindProtobufFieldInCatalystError
-
protobufFieldMatchError
-
unableToLocateProtobufMessageError
-
foundRecursionInProtobufSchema
-
protobufFieldTypeMismatchError
-
protobufClassLoadError
-
protobufDescriptorDependencyError
-
invalidByteStringFormatError
-
funcBuildError
-
ambiguousRelationAliasNameInNestedCTEError
-
ambiguousLateralColumnAliasError
-
ambiguousLateralColumnAliasError
-
lateralColumnAliasInAggFuncUnsupportedError
-
lateralColumnAliasInWindowUnsupportedError
-
lateralColumnAliasInAggWithWindowAndHavingUnsupportedError
-
dataTypeOperationUnsupportedError
-
nullableRowIdError
public static Throwable nullableRowIdError(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.AttributeReference> nullableRowIdAttrs) -
cannotRenameTableAcrossSchemaError
-
avroIncompatibleReadError
-
optionMustBeLiteralString
-
optionMustBeConstant
-
toSQLId
-
toSQLId
-
toSQLStmt
-
toSQLConf
-
toSQLType
-
toSQLType
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
toSQLValue
-
getSummary
-
getQueryContext
public static QueryContext[] getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) -
toSQLConfVal
-
toDSOption
-
toSQLExpr
-
toSQLSchema
-
toSQLValue
-
columnNotFoundError
-
descriptorParseError
-
cannotFindDescriptorFileError
-