public class DataTypeErrors
extends Object
QueryCompilationErrors
.Constructor and Description |
---|
DataTypeErrors() |
Modifier and Type | Method and Description |
---|---|
static Throwable |
ambiguousColumnOrFieldError(scala.collection.Seq<String> name,
int numMatches,
org.apache.spark.sql.catalyst.trees.Origin context) |
static Throwable |
attributeNameSyntaxError(String name) |
static ArithmeticException |
cannotChangeDecimalPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
cannotLoadUserDefinedTypeError(String name,
String userClass) |
static Throwable |
cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale,
int rightScale) |
static Throwable |
cannotMergeIncompatibleDataTypesError(DataType left,
DataType right) |
static ArithmeticException |
castingCauseOverflowError(String t,
DataType from,
DataType to) |
static Throwable |
charOrVarcharTypeAsStringUnsupportedError() |
static Throwable |
dataTypeUnsupportedError(String dataType,
String failure) |
static Throwable |
decimalCannotGreaterThanPrecisionError(int scale,
int precision) |
static org.apache.spark.SparkArithmeticException |
decimalPrecisionExceedsMaxPrecisionError(int precision,
int maxPrecision) |
static org.apache.spark.SparkRuntimeException |
failedParsingStructTypeError(String raw) |
static org.apache.spark.SparkUnsupportedOperationException |
fieldIndexOnRowWithoutSchemaError() |
static QueryContext[] |
getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static String |
getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext) |
static Throwable |
invalidDayTimeField(byte field,
scala.collection.Seq<String> supportedIds) |
static Throwable |
invalidDayTimeIntervalType(String startFieldName,
String endFieldName) |
static Throwable |
invalidFieldName(scala.collection.Seq<String> fieldName,
scala.collection.Seq<String> path,
org.apache.spark.sql.catalyst.trees.Origin context) |
static org.apache.spark.SparkNumberFormatException |
invalidInputInCastToNumberError(DataType to,
org.apache.spark.unsafe.types.UTF8String s,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static Throwable |
invalidYearMonthField(byte field,
scala.collection.Seq<String> supportedIds) |
static Throwable |
negativeScaleNotAllowedError(int scale) |
static Throwable |
notUserDefinedTypeError(String name,
String userClass) |
static org.apache.spark.SparkUnsupportedOperationException |
nullLiteralsCannotBeCastedError(String name) |
static org.apache.spark.SparkArithmeticException |
outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str) |
static Throwable |
schemaFailToParseError(String schema,
Throwable e) |
static String |
toSQLConf(String conf) |
static String |
toSQLId(scala.collection.Seq<String> parts) |
static String |
toSQLId(String parts) |
static String |
toSQLStmt(String text) |
static String |
toSQLType(org.apache.spark.sql.types.AbstractDataType t) |
static String |
toSQLType(String text) |
static String |
toSQLValue(double value) |
static String |
toSQLValue(float value) |
static String |
toSQLValue(int value) |
static String |
toSQLValue(long value) |
static String |
toSQLValue(short value) |
static String |
toSQLValue(String value) |
static String |
toSQLValue(org.apache.spark.unsafe.types.UTF8String value) |
static ArithmeticException |
unscaledValueTooLargeForPrecisionError(Decimal value,
int decimalPrecision,
int decimalScale,
org.apache.spark.sql.catalyst.trees.SQLQueryContext context) |
static org.apache.spark.SparkRuntimeException |
unsupportedArrayTypeError(Class<?> clazz) |
static org.apache.spark.SparkRuntimeException |
unsupportedJavaTypeError(Class<?> clazz) |
static org.apache.spark.SparkUnsupportedOperationException |
unsupportedOperationExceptionError() |
static SparkException |
unsupportedRoundingMode(scala.Enumeration.Value roundMode) |
static Throwable |
userSpecifiedSchemaUnsupportedError(String operation) |
static Throwable |
valueIsNullError(int index) |
public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationExceptionError()
public static org.apache.spark.SparkArithmeticException decimalPrecisionExceedsMaxPrecisionError(int precision, int maxPrecision)
public static SparkException unsupportedRoundingMode(scala.Enumeration.Value roundMode)
public static org.apache.spark.SparkArithmeticException outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str)
public static org.apache.spark.SparkRuntimeException unsupportedJavaTypeError(Class<?> clazz)
public static org.apache.spark.SparkUnsupportedOperationException nullLiteralsCannotBeCastedError(String name)
public static Throwable notUserDefinedTypeError(String name, String userClass)
public static Throwable cannotLoadUserDefinedTypeError(String name, String userClass)
public static org.apache.spark.SparkRuntimeException unsupportedArrayTypeError(Class<?> clazz)
public static Throwable schemaFailToParseError(String schema, Throwable e)
public static Throwable invalidDayTimeIntervalType(String startFieldName, String endFieldName)
public static Throwable invalidDayTimeField(byte field, scala.collection.Seq<String> supportedIds)
public static Throwable invalidYearMonthField(byte field, scala.collection.Seq<String> supportedIds)
public static Throwable decimalCannotGreaterThanPrecisionError(int scale, int precision)
public static Throwable negativeScaleNotAllowedError(int scale)
public static Throwable attributeNameSyntaxError(String name)
public static Throwable cannotMergeIncompatibleDataTypesError(DataType left, DataType right)
public static Throwable cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale, int rightScale)
public static Throwable dataTypeUnsupportedError(String dataType, String failure)
public static Throwable invalidFieldName(scala.collection.Seq<String> fieldName, scala.collection.Seq<String> path, org.apache.spark.sql.catalyst.trees.Origin context)
public static ArithmeticException unscaledValueTooLargeForPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
public static Throwable ambiguousColumnOrFieldError(scala.collection.Seq<String> name, int numMatches, org.apache.spark.sql.catalyst.trees.Origin context)
public static ArithmeticException castingCauseOverflowError(String t, DataType from, DataType to)
public static org.apache.spark.SparkRuntimeException failedParsingStructTypeError(String raw)
public static org.apache.spark.SparkUnsupportedOperationException fieldIndexOnRowWithoutSchemaError()
public static Throwable valueIsNullError(int index)
public static Throwable charOrVarcharTypeAsStringUnsupportedError()
public static Throwable userSpecifiedSchemaUnsupportedError(String operation)
public static String toSQLId(String parts)
public static String toSQLId(scala.collection.Seq<String> parts)
public static String toSQLStmt(String text)
public static String toSQLConf(String conf)
public static String toSQLType(String text)
public static String toSQLType(org.apache.spark.sql.types.AbstractDataType t)
public static String toSQLValue(String value)
public static String toSQLValue(org.apache.spark.unsafe.types.UTF8String value)
public static String toSQLValue(short value)
public static String toSQLValue(int value)
public static String toSQLValue(long value)
public static String toSQLValue(float value)
public static String toSQLValue(double value)
public static String getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)
public static QueryContext[] getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)