Key Features
- Customizable Chunking: Recursive, hierarchical, or fixed strategies
- Structured Data Support: Process CSV and Excel files
- Multiple Embedding Providers: Azure OpenAI, Mistral, Telekom OTC
- AI-Powered Retrieval: Vector-based semantic search
- Milvus Processing: Production-ready vector database engine
- Flexible Data Selection: Combine collections and individual data sources
- Status Monitoring: Track processing progress with detailed metadata
- Agent Integration: Direct use in agent configurations
Authentication
All endpoints require authentication using your API key:- API Key:
x-api-key: <key>
Available Endpoints
List Knowledge Bases
Retrieve all knowledge bases
Create Knowledge Base
Create a new knowledge base
Get Knowledge Base
Retrieve knowledge base details
Update Knowledge Base
Modify knowledge base settings
Delete Knowledge Base
Remove a knowledge base
Check Status
Monitor processing status
Available Collections
List collections ready for use
Processing Engine
Milvus Vector Database- High-performance vector similarity search
- Scalable for large document collections
- Advanced HNSW indexing algorithms
- Real-time updates and queries
- Production-ready for enterprise deployments
Milvus is currently the only supported processing engine. Additional engines may be added in future releases.
Knowledge Base Lifecycle
1. Creation
Define knowledge base with data sources, chunking strategy, and embedding configuration:2. Processing
Knowledge base processes data source content in two phases: Phase 1: Chunking (0-20% progress)- Extract text from documents (PDF, DOCX, TXT, MD, LOG)
- Process structured data (CSV, Excel) with custom row batching
- Apply chunking strategy to split into searchable chunks
- Track per-document progress
- Generate embeddings via batch API (Mistral, Azure OpenAI, or Telekom OTC)
- Stream embeddings to Milvus vector database
- Build HNSW indices for fast similarity search
- Validate data quality and store metadata
3. Completion
Ready knowledge bases can be used by agents for retrieval. Checkstatus, progress, and success_rate to monitor health.
Knowledge Base Status
pending
pending
Knowledge base created but processing not started
running
running
Collections are being processed and indexed
completed
completed
All collections processed and ready for queries
failed
failed
Processing failed - check error details and logs
Response Schema
Knowledge base responses include comprehensive processing details:Array of data source objects included in this knowledge base
Total number of data sources
Successfully processed data sources with full metadata
Data sources that failed processing with error details
Map of datasource_id to error messages for failed sources
Snapshot of datasource details at ingestion time with ingestion status
Percentage of successfully processed data sources (0-100)
Applied chunking configuration (recursive, hierarchical, or fixed)
Embedding model configuration (provider, model, dimensions)
Processing progress details including current phase and document being processed
Supported File Types
Documents:- PDF, DOC, DOCX
- TXT, MD (Markdown), LOG
- CSV (row-level or file-level processing)
- XLSX, XLS (Excel with batch row processing)
Chunking Strategies
Three strategies available for document processing:Recursive
Splits text using multiple separators in order. Best for general documents.- Configurable separators (paragraphs, sentences, words)
- Flexible chunk sizes and overlap
- Supports word or character splitting only
Hierarchical
Creates multi-level chunks preserving document structure. Ideal for research papers.- Multiple block sizes in descending order
- Maintains document hierarchy
- Word splitting only
Fixed
Simple fixed-size chunks. Fastest processing.- Consistent chunk sizes
- Configurable overlap
- Word splitting only
Embedding Providers
Azure OpenAI- text-embedding-ada-002 (1536 dimensions)
- mistral-embed (1024 dimensions)
- text-embedding-bge-m3 (1024 dimensions)
- jina-embeddings-v2-base-de (768 dimensions) - German-optimized
- jina-embeddings-v2-base-code (768 dimensions) - Code-optimized
- tsi-embedding-colqwen2-2b-v1 (1024 dimensions)
Agent Integration
Use knowledge bases in agent configurations:Retrieval Configuration
- Search parameters: Control relevance and results
- Context integration: How retrieved content is used
- Fallback behavior: When no relevant content is found
Performance Considerations
Indexing Time
- Collection size: Larger collections take longer to process
- Content complexity: Rich documents require more processing
- Engine choice: Different engines have varying performance characteristics
Query Performance
- Index optimization: Proper indexing improves search speed
- Result filtering: Limit results for faster responses
- Caching: Frequently accessed content is cached
Scalability
- Concurrent queries: Multiple agents can query simultaneously
- Update frequency: How often content changes
- Storage requirements: Vector storage grows with content
Monitoring and Troubleshooting
Health Checks
- Processing progress: Track completion percentage
- Error rates: Monitor failed data sources
- Query performance: Response times and accuracy
Common Issues
- Collection dependencies: Ensure collections are completed
- Processing failures: Check data source formats and content
- Memory limits: Large knowledge bases may hit resource limits
Use the status endpoint to monitor knowledge base health and processing progress.