Context Retrieval
Let Orama decide how to retrieve context on each message. Full-text, vector and hybrid search are built-in into Orama.
Custom Training
Train Orama in just a few shots. Customize and A/B test the system prompts. Visualize quality and speed performances.
Run Async JavaScript
Run async JavaScript hooks to enrich the context with information coming from any third-party system.
A Complete RAG Pipeline
Orama gives you a fully orchestrated retrieval-augmented generation flow. Automatically.
Every step of the pipeline is visual, transparent, and customizable: from interpreting the user’s query, to running full-text, vector, or hybrid search, to optimizing filters, merging results, and generating the final answer.
The pipeline adapts intelligently to your data and your configuration. You can plug in JavaScript hooks at any stage, fine-tune how retrieval works, or simply rely on Orama’s optimized defaults for fast, production-ready results.
Whether you're powering search, chat, or complex AI workflows, Orama handles all the logic behind the scenes, so you get reliable, explainable responses without managing any of the complexity.
Choose your LLM - or use an offline one.
Choose which LLM to use for each interaction. Orama connects to all the principal LLM providers out of the box. Need more data privacy? Use an offline model - no data will be shared with OpenAI, Anthropic, or any other company.
To maximize performance, embedding models are always offline.









