Class SnowflakeMetadataHandler
- java.lang.Object
-
- com.amazonaws.athena.connector.lambda.handlers.MetadataHandler
-
- com.amazonaws.athena.connectors.jdbc.manager.JdbcMetadataHandler
-
- com.amazonaws.athena.connectors.snowflake.SnowflakeMetadataHandler
-
- All Implemented Interfaces:
com.amazonaws.services.lambda.runtime.RequestStreamHandler
public class SnowflakeMetadataHandler extends JdbcMetadataHandler
Handles metadata for Snowflake. User must have access to `schemata`, `tables`, `columns` in information_schema.
-
-
Field Summary
-
Fields inherited from class com.amazonaws.athena.connectors.jdbc.manager.JdbcMetadataHandler
jdbcQueryPassthrough, TABLES_AND_VIEWS
-
Fields inherited from class com.amazonaws.athena.connector.lambda.handlers.MetadataHandler
configOptions, DISABLE_SPILL_ENCRYPTION, KMS_KEY_ID_ENV, SPILL_BUCKET_ENV, SPILL_PREFIX_ENV
-
-
Constructor Summary
Constructors Modifier Constructor Description SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, JdbcConnectionFactory jdbcConnectionFactory, Map<String,String> configOptions)
SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, Map<String,String> configOptions)
Used by Mux.protected
SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, software.amazon.awssdk.services.secretsmanager.SecretsManagerClient secretsManager, software.amazon.awssdk.services.athena.AthenaClient athena, JdbcConnectionFactory jdbcConnectionFactory, Map<String,String> configOptions)
SnowflakeMetadataHandler(Map<String,String> configOptions)
Instantiates handler to be used by Lambda function directly.
-
Method Summary
All Methods Static Methods Instance Methods Concrete Methods Modifier and Type Method Description GetDataSourceCapabilitiesResponse
doGetDataSourceCapabilities(BlockAllocator allocator, GetDataSourceCapabilitiesRequest request)
Used to describe the types of capabilities supported by a data source.GetSplitsResponse
doGetSplits(BlockAllocator blockAllocator, GetSplitsRequest getSplitsRequest)
Used to split-up the reads required to scan the requested batch of partition(s).GetTableResponse
doGetTable(BlockAllocator blockAllocator, GetTableRequest getTableRequest)
Used to get definition (field names, types, descriptions, etc...) of a Table.ListSchemasResponse
doListSchemaNames(BlockAllocator blockAllocator, ListSchemasRequest listSchemasRequest)
Used to get the list of schemas (aka databases) that this source contains.protected TableName
findTableNameFromQueryHint(TableName table)
Finding table name from query hint In sap hana schemas and tables can be case sensitive, but executed query from athena sends table and schema names in lower case, this has been handled by appending query hint to the table name as below "lambda:lambdaname".SCHEMA_NAME."void
getPartitions(BlockWriter blockWriter, GetTableLayoutRequest getTableLayoutRequest, QueryStatusChecker queryStatusChecker)
Snowflake manual partition logic based upon number of recordsorg.apache.arrow.vector.types.pojo.Schema
getPartitionSchema(String catalogName)
Delegates creation of partition schema to database type implementation.org.apache.arrow.vector.types.pojo.Schema
getSchema(Connection jdbcConnection, TableName tableName, org.apache.arrow.vector.types.pojo.Schema partitionSchema)
protected TableName
getTableFromMetadata(String catalogName, TableName tableHandle, DatabaseMetaData metadata)
Logic to handle case sensitivity of table name and schema nameprotected static Set<String>
listDatabaseNames(Connection jdbcConnection)
-
Methods inherited from class com.amazonaws.athena.connectors.jdbc.manager.JdbcMetadataHandler
caseInsensitiveTableSearch, convertDatasourceTypeToArrow, doGetQueryPassthroughSchema, doListTables, escapeNamePattern, getArrayArrowTypeFromTypeName, getCredentialProvider, getJdbcConnectionFactory, getSplitClauses, listPaginatedTables, listTables, setupQueryPassthroughSplit
-
Methods inherited from class com.amazonaws.athena.connector.lambda.handlers.MetadataHandler
doGetTableLayout, doHandleRequest, doPing, enhancePartitionSchema, getSecret, handleRequest, makeEncryptionKey, makeSpillLocation, onPing, resolveSecrets
-
-
-
-
Constructor Detail
-
SnowflakeMetadataHandler
public SnowflakeMetadataHandler(Map<String,String> configOptions)
Instantiates handler to be used by Lambda function directly. Recommend usingSnowflakeMuxCompositeHandler
instead.
-
SnowflakeMetadataHandler
public SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, Map<String,String> configOptions)
Used by Mux.
-
SnowflakeMetadataHandler
protected SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, software.amazon.awssdk.services.secretsmanager.SecretsManagerClient secretsManager, software.amazon.awssdk.services.athena.AthenaClient athena, JdbcConnectionFactory jdbcConnectionFactory, Map<String,String> configOptions)
-
SnowflakeMetadataHandler
public SnowflakeMetadataHandler(DatabaseConnectionConfig databaseConnectionConfig, JdbcConnectionFactory jdbcConnectionFactory, Map<String,String> configOptions)
-
-
Method Detail
-
doGetDataSourceCapabilities
public GetDataSourceCapabilitiesResponse doGetDataSourceCapabilities(BlockAllocator allocator, GetDataSourceCapabilitiesRequest request)
Description copied from class:MetadataHandler
Used to describe the types of capabilities supported by a data source. An engine can use this to determine what portions of the query to push down. A connector that returns any optimization will guarantee that the associated predicate will be pushed down.- Overrides:
doGetDataSourceCapabilities
in classMetadataHandler
- Parameters:
allocator
- Tool for creating and managing Apache Arrow Blocks.request
- Provides details about the catalog being used.- Returns:
- A GetDataSourceCapabilitiesResponse object which returns a map of supported optimizations that the connector is advertising to the consumer. The connector assumes all responsibility for whatever is passed here.
-
getPartitionSchema
public org.apache.arrow.vector.types.pojo.Schema getPartitionSchema(String catalogName)
Description copied from class:JdbcMetadataHandler
Delegates creation of partition schema to database type implementation.- Specified by:
getPartitionSchema
in classJdbcMetadataHandler
- Parameters:
catalogName
- Athena provided catalog name.- Returns:
- schema. See
Schema
-
getPartitions
public void getPartitions(BlockWriter blockWriter, GetTableLayoutRequest getTableLayoutRequest, QueryStatusChecker queryStatusChecker) throws Exception
Snowflake manual partition logic based upon number of records- Specified by:
getPartitions
in classJdbcMetadataHandler
- Parameters:
blockWriter
-getTableLayoutRequest
-queryStatusChecker
-- Throws:
Exception
-
doGetSplits
public GetSplitsResponse doGetSplits(BlockAllocator blockAllocator, GetSplitsRequest getSplitsRequest)
Description copied from class:MetadataHandler
Used to split-up the reads required to scan the requested batch of partition(s).- Specified by:
doGetSplits
in classJdbcMetadataHandler
- Parameters:
blockAllocator
- Tool for creating and managing Apache Arrow Blocks.getSplitsRequest
- Provides details of the catalog, database, table, andpartition(s) being queried as well as any filter predicate.- Returns:
- A GetSplitsResponse which primarily contains:
1. A Set
which represent read operations Amazon Athena must perform by calling your read function. 2. (Optional) A continuation token which allows you to paginate the generation of splits for large queries.
-
doGetTable
public GetTableResponse doGetTable(BlockAllocator blockAllocator, GetTableRequest getTableRequest) throws Exception
Description copied from class:MetadataHandler
Used to get definition (field names, types, descriptions, etc...) of a Table.- Overrides:
doGetTable
in classJdbcMetadataHandler
- Parameters:
blockAllocator
- Tool for creating and managing Apache Arrow Blocks.getTableRequest
- Provides details on who made the request and which Athena catalog, database, and table they are querying.- Returns:
- A GetTableResponse which primarily contains:
1. An Apache Arrow Schema object describing the table's columns, types, and descriptions.
2. A Set
of partition column names (or empty if the table isn't partitioned). - Throws:
Exception
-
getSchema
public org.apache.arrow.vector.types.pojo.Schema getSchema(Connection jdbcConnection, TableName tableName, org.apache.arrow.vector.types.pojo.Schema partitionSchema) throws Exception
- Parameters:
jdbcConnection
-tableName
-partitionSchema
-- Returns:
- Throws:
Exception
-
findTableNameFromQueryHint
protected TableName findTableNameFromQueryHint(TableName table)
Finding table name from query hint In sap hana schemas and tables can be case sensitive, but executed query from athena sends table and schema names in lower case, this has been handled by appending query hint to the table name as below "lambda:lambdaname".SCHEMA_NAME."TABLE_NAME@schemacase=upper&tablecase=upper"- Parameters:
table
-- Returns:
-
getTableFromMetadata
protected TableName getTableFromMetadata(String catalogName, TableName tableHandle, DatabaseMetaData metadata) throws SQLException
Logic to handle case sensitivity of table name and schema name- Parameters:
catalogName
-tableHandle
-metadata
-- Returns:
- Throws:
SQLException
-
doListSchemaNames
public ListSchemasResponse doListSchemaNames(BlockAllocator blockAllocator, ListSchemasRequest listSchemasRequest) throws Exception
Description copied from class:MetadataHandler
Used to get the list of schemas (aka databases) that this source contains.- Overrides:
doListSchemaNames
in classJdbcMetadataHandler
- Parameters:
blockAllocator
- Tool for creating and managing Apache Arrow Blocks.listSchemasRequest
- Provides details on who made the request and which Athena catalog they are querying.- Returns:
- A ListSchemasResponse which primarily contains a Set
of schema names and a catalog name corresponding the Athena catalog that was queried. - Throws:
Exception
-
listDatabaseNames
protected static Set<String> listDatabaseNames(Connection jdbcConnection) throws Exception
- Throws:
Exception
-
-