2. What it is
Knowledge Sources is the ingestion control layer for website URLs and documents. Every source is tracked with type, status, chunk count, retrieval count, and update timestamps.
Knowledge Sources
Manage URL and document sources with crawl mode control, source health visibility, and retrieval diagnostics.
Choose between indexing a single page or crawling a full documentation section depending on your rollout risk and breadth requirements. Single-page mode gives you precise control over what enters the knowledge base, while deep crawl mode ingests entire doc trees in one operation. Both modes provide real-time status visibility so teams always know where indexing stands.
Add one URL at a time to carefully validate content quality before expanding coverage to broader sections.
Crawl entire documentation sections automatically, capturing all linked pages in a single operation for fast initial setup.
Monitor each source as it moves through pending, fetching, and completed states so you can catch issues immediately.
Upload documents for parsing, chunking, and embedding so the AI assistant can retrieve knowledge from non-web assets like PDFs and internal guides. The ingestion pipeline automatically splits documents into optimally sized chunks and generates embeddings for semantic search. Each uploaded document is tracked as a document-type source with full lifecycle management.
Upload files through a dedicated interface that extracts text content and prepares it for knowledge base indexing.
Documents are split into retrieval-optimized chunks and embedded automatically, requiring no manual configuration.
Uploaded documents are tracked with the same status, metrics, and controls as crawled web sources for unified management.
Inspect source health and retrieval performance with detailed ranking diagnostics to improve answer grounding quality over time. Each source displays its current status, chunk counts, and retrieval metrics so teams can identify underperforming content. Recrawl controls with cooldown protection let you refresh stale sources without accidentally overwhelming the ingestion pipeline.
Every source shows its current lifecycle state so teams can quickly identify and resolve ingestion failures.
Refresh outdated sources on demand while built-in cooldown timers prevent accidental rapid repeated crawls.
See how often each source is retrieved and how well it ranks, so you can improve or replace low-performing content.
Knowledge Sources is the ingestion control layer for website URLs and documents. Every source is tracked with type, status, chunk count, retrieval count, and update timestamps.
Use this matrix to choose the right knowledge option for each support intent.
| Intent | Best option | Why | Link |
|---|---|---|---|
| Exact policy wording | Custom Answers | Deterministic response for high-risk phrasing. | /custom-answers |
| Short operational facts | Snippets | Fast to publish and update. | /snippets |
| Multi-step guides | Knowledge Pages | Better for long-form structure and context. | /knowledge-pages |
| Synced external docs | Knowledge Sources | Best for URL crawl and document ingestion workflows. | /knowledge-sources |
| Baseline product framing | Product Profile | Keeps answers aligned with core positioning. | /product-profile |
| Operational constraints and fallback rules | Assistant Contexts | Adds scoped behavioral context without rewriting docs. | /assistant-contexts |
FAQ
Use single-page for controlled experiments or sensitive pages. Use deep crawl for broad docs ingestion when structure is stable. Learn more
Yes. Recrawl is available per source, with cooldown controls to prevent rapid repeated runs. Learn more
Source detail views include retrieval diagnostics such as match counts, rank quality, and recent retrieval events. Learn more
Unify grounded AI, web and WhatsApp coverage, human handoff, and conversation analytics without stitching together separate tools.