Skip to main content

RAG (Retrieval-Augmented Generation)

Enhance LLM responses by connecting them to your own knowledge base, delivering accurate, context-specific answers without retraining.

Knowledge BaseChat AssistantVector SearchEmbedding ModelsChunking & ParsingGenAI API

Quick Start

Explore RAG

Knowledge Base

Create knowledge bases, manage documents, and evaluate retrieval quality


Chat Assistant

Build and configure chat assistants with custom prompts and LLM models


Advanced Features

Configure embedding models, chunking strategies, and retrieval optimization


Billing & Plans

Billing & Credits

RAG billing covers three components: storage for knowledge base documents, retrieval (search & indexing), and GenAI API calls for embeddings and LLM inference.

View Billing Docs →

Storage

Knowledge base documents and vector embeddings billed at Rs 8/GB/month.

Retrieval

Vector search and chunk retrieval charged at Rs 10 per 1 million tokens.

GenAI API

Embedding model and LLM token usage billed per GenAI model pricing. No additional cost when using Model Endpoints.