Class H2Dialect

Object
org.apache.spark.sql.jdbc.H2Dialect

public class H2Dialect extends Object
  • Constructor Details

    • H2Dialect

      public H2Dialect()
  • Method Details

    • canHandle

      public static boolean canHandle(String url)
    • isSupportedFunction

      public static boolean isSupportedFunction(String funcName)
    • getCatalystType

      public static scala.Option<DataType> getCatalystType(int sqlType, String typeName, int size, MetadataBuilder md)
    • getJDBCType

      public static scala.Option<JdbcType> getJDBCType(DataType dt)
    • registerFunction

      public static UnboundFunction registerFunction(String name, UnboundFunction fn)
    • functions

      public static scala.collection.Seq<scala.Tuple2<String,UnboundFunction>> functions()
    • clearFunctions

      public static void clearFunctions()
    • createIndex

      public static String createIndex(String indexName, Identifier tableIdent, NamedReference[] columns, Map<NamedReference,Map<String,String>> columnsProperties, Map<String,String> properties)
    • dropIndex

      public static String dropIndex(String indexName, Identifier tableIdent)
    • indexExists

      public static boolean indexExists(Connection conn, String indexName, Identifier tableIdent, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options)
    • listIndexes

      public static TableIndex[] listIndexes(Connection conn, Identifier tableIdent, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options)
    • classifyException

      public static AnalysisException classifyException(String message, Throwable e)
    • compileExpression

      public static scala.Option<String> compileExpression(Expression expr)
    • supportsLimit

      public static boolean supportsLimit()
    • supportsOffset

      public static boolean supportsOffset()
    • convertJavaTimestampToTimestamp

      public static Timestamp convertJavaTimestampToTimestamp(Timestamp t)
    • convertJavaTimestampToTimestampNTZ

      public static LocalDateTime convertJavaTimestampToTimestampNTZ(Timestamp t)
    • convertTimestampNTZToJavaTimestamp

      public static Timestamp convertTimestampNTZToJavaTimestamp(LocalDateTime ldt)
    • createConnectionFactory

      public static scala.Function1<Object,Connection> createConnectionFactory(org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options)
    • quoteIdentifier

      public static String quoteIdentifier(String colName)
    • createTable

      public static void createTable(Statement statement, String tableName, String strSchema, org.apache.spark.sql.execution.datasources.jdbc.JdbcOptionsInWrite options)
    • getTableExistsQuery

      public static String getTableExistsQuery(String table)
    • getSchemaQuery

      public static String getSchemaQuery(String table)
    • getTruncateQuery

      public static String getTruncateQuery(String table)
    • getTruncateQuery

      public static String getTruncateQuery(String table, scala.Option<Object> cascade)
    • getTruncateQuery$default$2

      public static scala.Option<Object> getTruncateQuery$default$2()
    • beforeFetch

      public static void beforeFetch(Connection connection, scala.collection.immutable.Map<String,String> properties)
    • compileValue

      public static Object compileValue(Object value)
    • compileAggregate

      public static scala.Option<String> compileAggregate(AggregateFunc aggFunction)
    • createSchema

      public static void createSchema(Statement statement, String schema, String comment)
    • schemasExists

      public static boolean schemasExists(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options, String schema)
    • listSchemas

      public static String[][] listSchemas(Connection conn, org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options)
    • isCascadingTruncateTable

      public static scala.Option<Object> isCascadingTruncateTable()
    • renameTable

      public static String renameTable(String oldTable, String newTable)
    • renameTable

      public static String renameTable(Identifier oldTable, Identifier newTable)
    • alterTable

      public static String[] alterTable(String tableName, scala.collection.Seq<TableChange> changes, int dbMajorVersion)
    • getAddColumnQuery

      public static String getAddColumnQuery(String tableName, String columnName, String dataType)
    • getRenameColumnQuery

      public static String getRenameColumnQuery(String tableName, String columnName, String newName, int dbMajorVersion)
    • getDeleteColumnQuery

      public static String getDeleteColumnQuery(String tableName, String columnName)
    • getUpdateColumnTypeQuery

      public static String getUpdateColumnTypeQuery(String tableName, String columnName, String newDataType)
    • getUpdateColumnNullabilityQuery

      public static String getUpdateColumnNullabilityQuery(String tableName, String columnName, boolean isNullable)
    • getTableCommentQuery

      public static String getTableCommentQuery(String table, String comment)
    • getSchemaCommentQuery

      public static String getSchemaCommentQuery(String schema, String comment)
    • removeSchemaCommentQuery

      public static String removeSchemaCommentQuery(String schema)
    • dropSchema

      public static String dropSchema(String schema, boolean cascade)
    • getLimitClause

      public static String getLimitClause(Integer limit)
    • getOffsetClause

      public static String getOffsetClause(Integer offset)
    • getJdbcSQLQueryBuilder

      public static JdbcSQLQueryBuilder getJdbcSQLQueryBuilder(org.apache.spark.sql.execution.datasources.jdbc.JDBCOptions options)
    • supportsTableSample

      public static boolean supportsTableSample()
    • getTableSample

      public static String getTableSample(org.apache.spark.sql.execution.datasources.v2.TableSampleInfo sample)
    • getFullyQualifiedQuotedTableName

      public static String getFullyQualifiedQuotedTableName(Identifier ident)
    • org$apache$spark$internal$Logging$$log_

      public static org.slf4j.Logger org$apache$spark$internal$Logging$$log_()
    • org$apache$spark$internal$Logging$$log__$eq

      public static void org$apache$spark$internal$Logging$$log__$eq(org.slf4j.Logger x$1)