Knowledge Source Configuration
Configure RAG: chunking profiles, embedding profiles, vector indexes, and knowledge sources. Ingest from upload, S3, or web.
Portal
Dashboard → Knowledge — Manage chunking profiles, embedding profiles, vector indexes, knowledge bases, and sources. Add sources (upload, S3, web), upload documents, trigger ingestion.
Setup flow
- Provider — Add a provider with embeddings (e.g. Voyage AI)
- Chunking profile — Strategy (fixed, markdown, semantic), chunk size, overlap
- Embedding profile — Provider model for embeddings
- Vector index — Backend (pgvector, Pinecone, Weaviate, Qdrant)
- Knowledge base — Container with default profiles + vector index
- Sources — Add upload, S3, or web source
- Ingestion — Upload docs, trigger ingestion job
Source types
- upload — Upload files via portal or API
- s3 — S3-compatible storage (MinIO, Scaleway, etc.)
- web — Fetch from URL
Pipeline
Ingestion: fetch documents → chunk → embed → index into vector store. Documents and chunks are stored in Postgres; embeddings go to the vector backend (pgvector, Pinecone, etc.).