Package org.apache.spark.sql.jdbc
Class DatabricksDialect
Object
org.apache.spark.sql.jdbc.JdbcDialect
org.apache.spark.sql.jdbc.DatabricksDialect
- All Implemented Interfaces:
Serializable
,org.apache.spark.internal.Logging
,scala.Equals
,scala.Product
- See Also:
-
Nested Class Summary
Nested classes/interfaces inherited from interface org.apache.spark.internal.Logging
org.apache.spark.internal.Logging.LogStringContext, org.apache.spark.internal.Logging.SparkShellLoggingFilter
-
Constructor Summary
-
Method Summary
Modifier and TypeMethodDescriptionabstract static R
apply()
boolean
Check if this dialect instance can handle a certain jdbc url.scala.Option<DataType>
getCatalystType
(int sqlType, String typeName, int size, MetadataBuilder md) Get the custom datatype mapping for the given jdbc meta information.scala.Option<JdbcType>
getJDBCType
(DataType dt) Retrieve the jdbc / sql type for a given datatype.getTableSample
(org.apache.spark.sql.execution.datasources.v2.TableSampleInfo sample) String[][]
listSchemas
(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options) Lists all the schemas in this table.quoteIdentifier
(String colName) Quotes the identifier.boolean
schemasExists
(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options, String schema) Check schema exists or not.boolean
Returns ture if dialect supports LIMIT clause.boolean
Returns ture if dialect supports OFFSET clause.boolean
static String
toString()
Methods inherited from class org.apache.spark.sql.jdbc.JdbcDialect
alterTable, beforeFetch, classifyException, classifyException, compileAggregate, compileExpression, compileValue, convertJavaDateToDate, convertJavaTimestampToTimestamp, convertJavaTimestampToTimestampNTZ, convertTimestampNTZToJavaTimestamp, createConnectionFactory, createIndex, createSchema, createTable, dropIndex, dropSchema, dropTable, functions, getAddColumnQuery, getDayTimeIntervalAsMicros, getDeleteColumnQuery, getFullyQualifiedQuotedTableName, getJdbcSQLQueryBuilder, getLimitClause, getOffsetClause, getRenameColumnQuery, getSchemaCommentQuery, getSchemaQuery, getTableCommentQuery, getTableExistsQuery, getTruncateQuery, getTruncateQuery, getUpdateColumnNullabilityQuery, getUpdateColumnTypeQuery, getYearMonthIntervalAsMonths, indexExists, insertIntoTable, isCascadingTruncateTable, isSupportedFunction, listIndexes, removeSchemaCommentQuery, renameTable, renameTable, updateExtraColumnMeta
Methods inherited from class java.lang.Object
equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
Methods inherited from interface scala.Equals
canEqual, equals
Methods inherited from interface org.apache.spark.internal.Logging
initializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logDebug, logDebug, logError, logError, logError, logError, logInfo, logInfo, logInfo, logInfo, logName, LogStringContext, logTrace, logTrace, logTrace, logTrace, logWarning, logWarning, logWarning, logWarning, org$apache$spark$internal$Logging$$log_, org$apache$spark$internal$Logging$$log__$eq, withLogContext
Methods inherited from interface scala.Product
productArity, productElement, productElementName, productElementNames, productIterator, productPrefix
-
Constructor Details
-
DatabricksDialect
public DatabricksDialect()
-
-
Method Details
-
apply
public abstract static R apply() -
toString
-
canHandle
Description copied from class:JdbcDialect
Check if this dialect instance can handle a certain jdbc url.- Specified by:
canHandle
in classJdbcDialect
- Parameters:
url
- the jdbc url.- Returns:
- True if the dialect can be applied on the given jdbc url.
-
getCatalystType
public scala.Option<DataType> getCatalystType(int sqlType, String typeName, int size, MetadataBuilder md) Description copied from class:JdbcDialect
Get the custom datatype mapping for the given jdbc meta information.Guidelines for mapping database defined timestamps to Spark SQL timestamps:
-
TIMESTAMP WITHOUT TIME ZONE if preferTimestampNTZ ->
TimestampNTZType
-
TIMESTAMP WITHOUT TIME ZONE if !preferTimestampNTZ ->
TimestampType
(LTZ) - TIMESTAMP WITH TIME ZONE ->
TimestampType
(LTZ) - TIMESTAMP WITH LOCAL TIME ZONE ->
TimestampType
(LTZ) -
If the TIMESTAMP cannot be distinguished by
sqlType
andtypeName
, preferTimestampNTZ is respected for now, but we may need to add another option in the future if necessary.
- Overrides:
getCatalystType
in classJdbcDialect
- Parameters:
sqlType
- Refers toTypes
constants, or other constants defined by the target database, e.g.-101
is Oracle's TIMESTAMP WITH TIME ZONE type. This value is returned byResultSetMetaData.getColumnType(int)
.typeName
- The column type name used by the database (e.g. "BIGINT UNSIGNED"). This is sometimes used to determine the target data type whensqlType
is not sufficient if multiple database types are conflated into a single id. This value is returned byResultSetMetaData.getColumnTypeName(int)
.size
- The size of the type, e.g. the maximum precision for numeric types, length for character string, etc. This value is returned byResultSetMetaData.getPrecision(int)
.md
- Result metadata associated with this type. This contains additional information fromResultSetMetaData
or user specified options.-
isTimestampNTZ
: Whether read a TIMESTAMP WITHOUT TIME ZONE value asTimestampNTZType
or not. This is configured byJDBCOptions.preferTimestampNTZ
. -
scale
: The length of fractional partResultSetMetaData.getScale(int)
-
- Returns:
- An option the actual DataType (subclasses of
DataType
) or None if the default type mapping should be used.
-
TIMESTAMP WITHOUT TIME ZONE if preferTimestampNTZ ->
-
getJDBCType
Description copied from class:JdbcDialect
Retrieve the jdbc / sql type for a given datatype.- Overrides:
getJDBCType
in classJdbcDialect
- Parameters:
dt
- The datatype (e.g.StringType
)- Returns:
- The new JdbcType if there is an override for this DataType
-
quoteIdentifier
Description copied from class:JdbcDialect
Quotes the identifier. This is used to put quotes around the identifier in case the column name is a reserved keyword, or in case it contains characters that require quotes (e.g. space).- Overrides:
quoteIdentifier
in classJdbcDialect
- Parameters:
colName
- (undocumented)- Returns:
- (undocumented)
-
supportsLimit
public boolean supportsLimit()Description copied from class:JdbcDialect
Returns ture if dialect supports LIMIT clause.Note: Some build-in dialect supports LIMIT clause with some trick, please see:
OracleDialect.OracleSQLQueryBuilder
andMsSqlServerDialect.MsSqlServerSQLQueryBuilder
.- Overrides:
supportsLimit
in classJdbcDialect
- Returns:
- (undocumented)
-
supportsOffset
public boolean supportsOffset()Description copied from class:JdbcDialect
Returns ture if dialect supports OFFSET clause.Note: Some build-in dialect supports OFFSET clause with some trick, please see:
OracleDialect.OracleSQLQueryBuilder
andMySQLDialect.MySQLSQLQueryBuilder
.- Overrides:
supportsOffset
in classJdbcDialect
- Returns:
- (undocumented)
-
supportsTableSample
public boolean supportsTableSample()- Overrides:
supportsTableSample
in classJdbcDialect
-
getTableSample
- Overrides:
getTableSample
in classJdbcDialect
-
schemasExists
public boolean schemasExists(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options, String schema) Description copied from class:JdbcDialect
Check schema exists or not.- Overrides:
schemasExists
in classJdbcDialect
- Parameters:
conn
- (undocumented)options
- (undocumented)schema
- (undocumented)- Returns:
- (undocumented)
-
listSchemas
public String[][] listSchemas(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options) Description copied from class:JdbcDialect
Lists all the schemas in this table.- Overrides:
listSchemas
in classJdbcDialect
- Parameters:
conn
- (undocumented)options
- (undocumented)- Returns:
- (undocumented)
-