We recommend new projects start with resources from the AWS provider.
aws-native.glue.getCrawler
Explore with Pulumi AI
We recommend new projects start with resources from the AWS provider.
Resource Type definition for AWS::Glue::Crawler
Using getCrawler
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getCrawler(args: GetCrawlerArgs, opts?: InvokeOptions): Promise<GetCrawlerResult>
function getCrawlerOutput(args: GetCrawlerOutputArgs, opts?: InvokeOptions): Output<GetCrawlerResult>def get_crawler(name: Optional[str] = None,
                opts: Optional[InvokeOptions] = None) -> GetCrawlerResult
def get_crawler_output(name: Optional[pulumi.Input[str]] = None,
                opts: Optional[InvokeOptions] = None) -> Output[GetCrawlerResult]func LookupCrawler(ctx *Context, args *LookupCrawlerArgs, opts ...InvokeOption) (*LookupCrawlerResult, error)
func LookupCrawlerOutput(ctx *Context, args *LookupCrawlerOutputArgs, opts ...InvokeOption) LookupCrawlerResultOutput> Note: This function is named LookupCrawler in the Go SDK.
public static class GetCrawler 
{
    public static Task<GetCrawlerResult> InvokeAsync(GetCrawlerArgs args, InvokeOptions? opts = null)
    public static Output<GetCrawlerResult> Invoke(GetCrawlerInvokeArgs args, InvokeOptions? opts = null)
}public static CompletableFuture<GetCrawlerResult> getCrawler(GetCrawlerArgs args, InvokeOptions options)
public static Output<GetCrawlerResult> getCrawler(GetCrawlerArgs args, InvokeOptions options)
fn::invoke:
  function: aws-native:glue:getCrawler
  arguments:
    # arguments dictionaryThe following arguments are supported:
- Name string
- The name of the crawler.
- Name string
- The name of the crawler.
- name String
- The name of the crawler.
- name string
- The name of the crawler.
- name str
- The name of the crawler.
- name String
- The name of the crawler.
getCrawler Result
The following output properties are available:
- Classifiers List<string>
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- Configuration string
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- CrawlerSecurity stringConfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- DatabaseName string
- The name of the database in which the crawler's output is stored.
- Description string
- A description of the crawler.
- LakeFormation Pulumi.Configuration Aws Native. Glue. Outputs. Crawler Lake Formation Configuration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- RecrawlPolicy Pulumi.Aws Native. Glue. Outputs. Crawler Recrawl Policy 
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- Role string
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- Schedule
Pulumi.Aws Native. Glue. Outputs. Crawler Schedule 
- For scheduled crawlers, the schedule when the crawler runs.
- SchemaChange Pulumi.Policy Aws Native. Glue. Outputs. Crawler Schema Change Policy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- TablePrefix string
- The prefix added to the names of tables that are created.
- object
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- Targets
Pulumi.Aws Native. Glue. Outputs. Crawler Targets 
- A collection of targets to crawl.
- Classifiers []string
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- Configuration string
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- CrawlerSecurity stringConfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- DatabaseName string
- The name of the database in which the crawler's output is stored.
- Description string
- A description of the crawler.
- LakeFormation CrawlerConfiguration Lake Formation Configuration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- RecrawlPolicy CrawlerRecrawl Policy 
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- Role string
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- Schedule
CrawlerSchedule 
- For scheduled crawlers, the schedule when the crawler runs.
- SchemaChange CrawlerPolicy Schema Change Policy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- TablePrefix string
- The prefix added to the names of tables that are created.
- interface{}
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- Targets
CrawlerTargets 
- A collection of targets to crawl.
- classifiers List<String>
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- configuration String
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- crawlerSecurity StringConfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- databaseName String
- The name of the database in which the crawler's output is stored.
- description String
- A description of the crawler.
- lakeFormation CrawlerConfiguration Lake Formation Configuration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- recrawlPolicy CrawlerRecrawl Policy 
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- role String
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- schedule
CrawlerSchedule 
- For scheduled crawlers, the schedule when the crawler runs.
- schemaChange CrawlerPolicy Schema Change Policy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- tablePrefix String
- The prefix added to the names of tables that are created.
- Object
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- targets
CrawlerTargets 
- A collection of targets to crawl.
- classifiers string[]
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- configuration string
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- crawlerSecurity stringConfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- databaseName string
- The name of the database in which the crawler's output is stored.
- description string
- A description of the crawler.
- lakeFormation CrawlerConfiguration Lake Formation Configuration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- recrawlPolicy CrawlerRecrawl Policy 
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- role string
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- schedule
CrawlerSchedule 
- For scheduled crawlers, the schedule when the crawler runs.
- schemaChange CrawlerPolicy Schema Change Policy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- tablePrefix string
- The prefix added to the names of tables that are created.
- any
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- targets
CrawlerTargets 
- A collection of targets to crawl.
- classifiers Sequence[str]
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- configuration str
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- crawler_security_ strconfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- database_name str
- The name of the database in which the crawler's output is stored.
- description str
- A description of the crawler.
- lake_formation_ Crawlerconfiguration Lake Formation Configuration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- recrawl_policy CrawlerRecrawl Policy 
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- role str
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- schedule
CrawlerSchedule 
- For scheduled crawlers, the schedule when the crawler runs.
- schema_change_ Crawlerpolicy Schema Change Policy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- table_prefix str
- The prefix added to the names of tables that are created.
- Any
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- targets
CrawlerTargets 
- A collection of targets to crawl.
- classifiers List<String>
- A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
- configuration String
- Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
- crawlerSecurity StringConfiguration 
- The name of the SecurityConfiguration structure to be used by this crawler.
- databaseName String
- The name of the database in which the crawler's output is stored.
- description String
- A description of the crawler.
- lakeFormation Property MapConfiguration 
- Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- recrawlPolicy Property Map
- A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
- role String
- The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
- schedule Property Map
- For scheduled crawlers, the schedule when the crawler runs.
- schemaChange Property MapPolicy 
- The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The - SchemaChangePolicydoes not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the- SchemaChangePolicyon a crawler.- The SchemaChangePolicy consists of two components, - UpdateBehaviorand- DeleteBehavior.
- tablePrefix String
- The prefix added to the names of tables that are created.
- Any
- The tags to use with this crawler. - Search the CloudFormation User Guide for - AWS::Glue::Crawlerfor more information about the expected schema for this property.
- targets Property Map
- A collection of targets to crawl.
Supporting Types
CrawlerCatalogTarget  
- ConnectionName string
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- DatabaseName string
- The name of the database to be synchronized.
- DlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- EventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- Tables List<string>
- A list of the tables to be synchronized.
- ConnectionName string
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- DatabaseName string
- The name of the database to be synchronized.
- DlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- EventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- Tables []string
- A list of the tables to be synchronized.
- connectionName String
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- databaseName String
- The name of the database to be synchronized.
- dlqEvent StringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue StringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- tables List<String>
- A list of the tables to be synchronized.
- connectionName string
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- databaseName string
- The name of the database to be synchronized.
- dlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- tables string[]
- A list of the tables to be synchronized.
- connection_name str
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- database_name str
- The name of the database to be synchronized.
- dlq_event_ strqueue_ arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- event_queue_ strarn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- tables Sequence[str]
- A list of the tables to be synchronized.
- connectionName String
- The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
- databaseName String
- The name of the database to be synchronized.
- dlqEvent StringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue StringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- tables List<String>
- A list of the tables to be synchronized.
CrawlerDeltaTarget  
- ConnectionName string
- The name of the connection to use to connect to the Delta table target.
- CreateNative boolDelta Table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- DeltaTables List<string>
- A list of the Amazon S3 paths to the Delta tables.
- WriteManifest bool
- Specifies whether to write the manifest files to the Delta table path.
- ConnectionName string
- The name of the connection to use to connect to the Delta table target.
- CreateNative boolDelta Table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- DeltaTables []string
- A list of the Amazon S3 paths to the Delta tables.
- WriteManifest bool
- Specifies whether to write the manifest files to the Delta table path.
- connectionName String
- The name of the connection to use to connect to the Delta table target.
- createNative BooleanDelta Table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- deltaTables List<String>
- A list of the Amazon S3 paths to the Delta tables.
- writeManifest Boolean
- Specifies whether to write the manifest files to the Delta table path.
- connectionName string
- The name of the connection to use to connect to the Delta table target.
- createNative booleanDelta Table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- deltaTables string[]
- A list of the Amazon S3 paths to the Delta tables.
- writeManifest boolean
- Specifies whether to write the manifest files to the Delta table path.
- connection_name str
- The name of the connection to use to connect to the Delta table target.
- create_native_ booldelta_ table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- delta_tables Sequence[str]
- A list of the Amazon S3 paths to the Delta tables.
- write_manifest bool
- Specifies whether to write the manifest files to the Delta table path.
- connectionName String
- The name of the connection to use to connect to the Delta table target.
- createNative BooleanDelta Table 
- Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
- deltaTables List<String>
- A list of the Amazon S3 paths to the Delta tables.
- writeManifest Boolean
- Specifies whether to write the manifest files to the Delta table path.
CrawlerDynamoDbTarget   
- Path string
- The name of the DynamoDB table to crawl.
- Path string
- The name of the DynamoDB table to crawl.
- path String
- The name of the DynamoDB table to crawl.
- path string
- The name of the DynamoDB table to crawl.
- path str
- The name of the DynamoDB table to crawl.
- path String
- The name of the DynamoDB table to crawl.
CrawlerHudiTarget  
- ConnectionName string
- The name of the connection to use to connect to the Hudi target.
- Exclusions List<string>
- A list of global patterns used to exclude from the crawl.
- MaximumTraversal intDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- Paths List<string>
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
- ConnectionName string
- The name of the connection to use to connect to the Hudi target.
- Exclusions []string
- A list of global patterns used to exclude from the crawl.
- MaximumTraversal intDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- Paths []string
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
- connectionName String
- The name of the connection to use to connect to the Hudi target.
- exclusions List<String>
- A list of global patterns used to exclude from the crawl.
- maximumTraversal IntegerDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths List<String>
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
- connectionName string
- The name of the connection to use to connect to the Hudi target.
- exclusions string[]
- A list of global patterns used to exclude from the crawl.
- maximumTraversal numberDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths string[]
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
- connection_name str
- The name of the connection to use to connect to the Hudi target.
- exclusions Sequence[str]
- A list of global patterns used to exclude from the crawl.
- maximum_traversal_ intdepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths Sequence[str]
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
- connectionName String
- The name of the connection to use to connect to the Hudi target.
- exclusions List<String>
- A list of global patterns used to exclude from the crawl.
- maximumTraversal NumberDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths List<String>
- One or more Amazon S3 paths that contains Hudi metadata folders as s3://bucket/prefix .
CrawlerIcebergTarget  
- ConnectionName string
- The name of the connection to use to connect to the Iceberg target.
- Exclusions List<string>
- A list of global patterns used to exclude from the crawl.
- MaximumTraversal intDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- Paths List<string>
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
- ConnectionName string
- The name of the connection to use to connect to the Iceberg target.
- Exclusions []string
- A list of global patterns used to exclude from the crawl.
- MaximumTraversal intDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- Paths []string
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
- connectionName String
- The name of the connection to use to connect to the Iceberg target.
- exclusions List<String>
- A list of global patterns used to exclude from the crawl.
- maximumTraversal IntegerDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths List<String>
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
- connectionName string
- The name of the connection to use to connect to the Iceberg target.
- exclusions string[]
- A list of global patterns used to exclude from the crawl.
- maximumTraversal numberDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths string[]
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
- connection_name str
- The name of the connection to use to connect to the Iceberg target.
- exclusions Sequence[str]
- A list of global patterns used to exclude from the crawl.
- maximum_traversal_ intdepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths Sequence[str]
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
- connectionName String
- The name of the connection to use to connect to the Iceberg target.
- exclusions List<String>
- A list of global patterns used to exclude from the crawl.
- maximumTraversal NumberDepth 
- The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
- paths List<String>
- One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
CrawlerJdbcTarget  
- ConnectionName string
- The name of the connection to use to connect to the JDBC target.
- EnableAdditional List<string>Metadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- Exclusions List<string>
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- Path string
- The path of the JDBC target.
- ConnectionName string
- The name of the connection to use to connect to the JDBC target.
- EnableAdditional []stringMetadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- Exclusions []string
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- Path string
- The path of the JDBC target.
- connectionName String
- The name of the connection to use to connect to the JDBC target.
- enableAdditional List<String>Metadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- exclusions List<String>
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- path String
- The path of the JDBC target.
- connectionName string
- The name of the connection to use to connect to the JDBC target.
- enableAdditional string[]Metadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- exclusions string[]
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- path string
- The path of the JDBC target.
- connection_name str
- The name of the connection to use to connect to the JDBC target.
- enable_additional_ Sequence[str]metadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- exclusions Sequence[str]
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- path str
- The path of the JDBC target.
- connectionName String
- The name of the connection to use to connect to the JDBC target.
- enableAdditional List<String>Metadata 
- Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database. - If you do not need additional metadata, keep the field empty. 
- exclusions List<String>
- A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
- path String
- The path of the JDBC target.
CrawlerLakeFormationConfiguration   
- AccountId string
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- UseLake boolFormation Credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- AccountId string
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- UseLake boolFormation Credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- accountId String
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- useLake BooleanFormation Credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- accountId string
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- useLake booleanFormation Credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- account_id str
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- use_lake_ boolformation_ credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
- accountId String
- Required for cross account crawls. For same account crawls as the target data, this can be left as null.
- useLake BooleanFormation Credentials 
- Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
CrawlerMongoDbTarget   
- ConnectionName string
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- Path string
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
- ConnectionName string
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- Path string
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
- connectionName String
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- path String
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
- connectionName string
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- path string
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
- connection_name str
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- path str
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
- connectionName String
- The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
- path String
- The path of the Amazon DocumentDB or MongoDB target (database/collection).
CrawlerRecrawlPolicy  
- RecrawlBehavior string
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
- RecrawlBehavior string
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
- recrawlBehavior String
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
- recrawlBehavior string
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
- recrawl_behavior str
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
- recrawlBehavior String
- Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
CrawlerS3Target 
- ConnectionName string
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- DlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- EventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- Exclusions List<string>
- A list of glob patterns used to exclude from the crawl.
- Path string
- The path to the Amazon S3 target.
- SampleSize int
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
- ConnectionName string
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- DlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- EventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- Exclusions []string
- A list of glob patterns used to exclude from the crawl.
- Path string
- The path to the Amazon S3 target.
- SampleSize int
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
- connectionName String
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- dlqEvent StringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue StringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- exclusions List<String>
- A list of glob patterns used to exclude from the crawl.
- path String
- The path to the Amazon S3 target.
- sampleSize Integer
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
- connectionName string
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- dlqEvent stringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue stringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- exclusions string[]
- A list of glob patterns used to exclude from the crawl.
- path string
- The path to the Amazon S3 target.
- sampleSize number
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
- connection_name str
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- dlq_event_ strqueue_ arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- event_queue_ strarn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- exclusions Sequence[str]
- A list of glob patterns used to exclude from the crawl.
- path str
- The path to the Amazon S3 target.
- sample_size int
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
- connectionName String
- The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
- dlqEvent StringQueue Arn 
- A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
- eventQueue StringArn 
- A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
- exclusions List<String>
- A list of glob patterns used to exclude from the crawl.
- path String
- The path to the Amazon S3 target.
- sampleSize Number
- Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
CrawlerSchedule 
- ScheduleExpression string
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
- ScheduleExpression string
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
- scheduleExpression String
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
- scheduleExpression string
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
- schedule_expression str
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
- scheduleExpression String
- A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
CrawlerSchemaChangePolicy   
- DeleteBehavior string
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- UpdateBehavior string
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
- DeleteBehavior string
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- UpdateBehavior string
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
- deleteBehavior String
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- updateBehavior String
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
- deleteBehavior string
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- updateBehavior string
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
- delete_behavior str
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- update_behavior str
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
- deleteBehavior String
- The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
- updateBehavior String
- The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
CrawlerTargets 
- CatalogTargets List<Pulumi.Aws Native. Glue. Inputs. Crawler Catalog Target> 
- Specifies AWS Glue Data Catalog targets.
- DeltaTargets List<Pulumi.Aws Native. Glue. Inputs. Crawler Delta Target> 
- Specifies an array of Delta data store targets.
- DynamoDb List<Pulumi.Targets Aws Native. Glue. Inputs. Crawler Dynamo Db Target> 
- Specifies Amazon DynamoDB targets.
- HudiTargets List<Pulumi.Aws Native. Glue. Inputs. Crawler Hudi Target> 
- Specifies Apache Hudi data store targets.
- IcebergTargets List<Pulumi.Aws Native. Glue. Inputs. Crawler Iceberg Target> 
- Specifies Apache Iceberg data store targets.
- JdbcTargets List<Pulumi.Aws Native. Glue. Inputs. Crawler Jdbc Target> 
- Specifies JDBC targets.
- MongoDb List<Pulumi.Targets Aws Native. Glue. Inputs. Crawler Mongo Db Target> 
- A list of Mongo DB targets.
- S3Targets
List<Pulumi.Aws Native. Glue. Inputs. Crawler S3Target> 
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
- CatalogTargets []CrawlerCatalog Target 
- Specifies AWS Glue Data Catalog targets.
- DeltaTargets []CrawlerDelta Target 
- Specifies an array of Delta data store targets.
- DynamoDb []CrawlerTargets Dynamo Db Target 
- Specifies Amazon DynamoDB targets.
- HudiTargets []CrawlerHudi Target 
- Specifies Apache Hudi data store targets.
- IcebergTargets []CrawlerIceberg Target 
- Specifies Apache Iceberg data store targets.
- JdbcTargets []CrawlerJdbc Target 
- Specifies JDBC targets.
- MongoDb []CrawlerTargets Mongo Db Target 
- A list of Mongo DB targets.
- S3Targets
[]CrawlerS3Target 
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
- catalogTargets List<CrawlerCatalog Target> 
- Specifies AWS Glue Data Catalog targets.
- deltaTargets List<CrawlerDelta Target> 
- Specifies an array of Delta data store targets.
- dynamoDb List<CrawlerTargets Dynamo Db Target> 
- Specifies Amazon DynamoDB targets.
- hudiTargets List<CrawlerHudi Target> 
- Specifies Apache Hudi data store targets.
- icebergTargets List<CrawlerIceberg Target> 
- Specifies Apache Iceberg data store targets.
- jdbcTargets List<CrawlerJdbc Target> 
- Specifies JDBC targets.
- mongoDb List<CrawlerTargets Mongo Db Target> 
- A list of Mongo DB targets.
- s3Targets
List<CrawlerS3Target> 
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
- catalogTargets CrawlerCatalog Target[] 
- Specifies AWS Glue Data Catalog targets.
- deltaTargets CrawlerDelta Target[] 
- Specifies an array of Delta data store targets.
- dynamoDb CrawlerTargets Dynamo Db Target[] 
- Specifies Amazon DynamoDB targets.
- hudiTargets CrawlerHudi Target[] 
- Specifies Apache Hudi data store targets.
- icebergTargets CrawlerIceberg Target[] 
- Specifies Apache Iceberg data store targets.
- jdbcTargets CrawlerJdbc Target[] 
- Specifies JDBC targets.
- mongoDb CrawlerTargets Mongo Db Target[] 
- A list of Mongo DB targets.
- s3Targets
CrawlerS3Target[] 
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
- catalog_targets Sequence[CrawlerCatalog Target] 
- Specifies AWS Glue Data Catalog targets.
- delta_targets Sequence[CrawlerDelta Target] 
- Specifies an array of Delta data store targets.
- dynamo_db_ Sequence[Crawlertargets Dynamo Db Target] 
- Specifies Amazon DynamoDB targets.
- hudi_targets Sequence[CrawlerHudi Target] 
- Specifies Apache Hudi data store targets.
- iceberg_targets Sequence[CrawlerIceberg Target] 
- Specifies Apache Iceberg data store targets.
- jdbc_targets Sequence[CrawlerJdbc Target] 
- Specifies JDBC targets.
- mongo_db_ Sequence[Crawlertargets Mongo Db Target] 
- A list of Mongo DB targets.
- s3_targets Sequence[CrawlerS3Target] 
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
- catalogTargets List<Property Map>
- Specifies AWS Glue Data Catalog targets.
- deltaTargets List<Property Map>
- Specifies an array of Delta data store targets.
- dynamoDb List<Property Map>Targets 
- Specifies Amazon DynamoDB targets.
- hudiTargets List<Property Map>
- Specifies Apache Hudi data store targets.
- icebergTargets List<Property Map>
- Specifies Apache Iceberg data store targets.
- jdbcTargets List<Property Map>
- Specifies JDBC targets.
- mongoDb List<Property Map>Targets 
- A list of Mongo DB targets.
- s3Targets List<Property Map>
- Specifies Amazon Simple Storage Service (Amazon S3) targets.
Package Details
- Repository
- AWS Native pulumi/pulumi-aws-native
- License
- Apache-2.0
We recommend new projects start with resources from the AWS provider.