@cdklabs/generative-ai-cdk-constructs
@cdklabs/generative-ai-cdk-constructs / bedrock / WebCrawlerDataSourceAssociationProps
Interface to add a new data source to an existing KB.
readonlyoptionalchunkingStrategy:ChunkingStrategy
The chunking stategy to use for splitting your documents or content. The chunks are then converted to embeddings and written to the vector index allowing for similarity search and retrieval of the content.
ChunkingStrategy.DEFAULT
DataSourceAssociationProps.chunkingStrategy
readonlyoptionalcontextEnrichment:ContextEnrichment
The context enrichment configuration to use.
- No context enrichment is used.
DataSourceAssociationProps.contextEnrichment
readonlyoptionalcrawlingRate:number
The max rate at which pages are crawled, up to 300 per minute per host. Higher values will decrease sync time but increase the load on the host.
300
readonlyoptionalcrawlingScope:CrawlingScope
The scope of the crawling.
- CrawlingScope.DEFAULT
readonlyoptionalcustomTransformation:CustomTransformation
The custom transformation strategy to use.
- No custom transformation is used.
DataSourceAssociationProps.customTransformation
readonlyoptionaldataDeletionPolicy:DataDeletionPolicy
The data deletion policy to apply to the data source.
- Sets the data deletion policy to the default of the data source type.
DataSourceAssociationProps.dataDeletionPolicy
readonlyoptionaldataSourceName:string
The name of the data source.
- A new name will be generated.
DataSourceAssociationProps.dataSourceName
readonlyoptionaldescription:string
A description of the data source.
- No description is provided.
DataSourceAssociationProps.description
readonlyoptionalfilters:CrawlingFilters
The filters (regular expression patterns) for the crawling. If there’s a conflict, the exclude pattern takes precedence.
None
readonlyoptionalkmsKey:IKey
The KMS key to use to encrypt the data source.
- Service owned and managed key.
DataSourceAssociationProps.kmsKey
readonlyoptionalmaxPages:number
The maximum number of pages to crawl. The max number of web pages crawled from your source URLs, up to 25,000 pages. If the web pages exceed this limit, the data source sync will fail and no web pages will be ingested.
- No limit
readonlyoptionalparsingStrategy:ParsingStrategy
The parsing strategy to use.
- No Parsing Stategy is used.
DataSourceAssociationProps.parsingStrategy
readonlysourceUrls:string[]
The source urls in the format https://www.sitename.com.
Maximum of 100 URLs.
readonlyoptionaluserAgent:string
The user agent string to use when crawling.
- Default user agent string
readonlyoptionaluserAgentHeader:string
The user agent header to use when crawling. A string used for identifying the crawler or bot when it accesses a web server. The user agent header value consists of the bedrockbot, UUID, and a user agent suffix for your crawler (if one is provided). By default, it is set to bedrockbot_UUID. You can optionally append a custom suffix to bedrockbot_UUID to allowlist a specific user agent permitted to access your source URLs.
- Default user agent header (bedrockbot_UUID)