Skip to main content
Conversational search is an AI-powered feature built on top of Meilisearch’s search engine. It works as a built-in Retrieval Augmented Generation (RAG) system: when a user asks a question, Meilisearch retrieves relevant documents from its indexes, then uses an LLM to generate a response grounded in those results. The AI never answers from its own general knowledge. Every response is based on the data you have indexed in Meilisearch. This is similar to how Perplexity works: every answer comes with source documents so users can verify the information. Meilisearch brings the same pattern to your own data.
Conversational search relies on large language models (LLMs) to generate responses. LLMs may occasionally hallucinate inaccurate or misleading information, even when provided with correct source documents. Always monitor responses in production environments and consider implementing guardrails to reduce this risk.

Use cases

Conversational search supports three main use cases, all powered by the same /chats API route:

Multi-turn chat

Build a full conversational interface where users ask follow-up questions and the agent maintains context across the conversation. This is ideal for knowledge bases, customer support, and documentation search. Example: A user asks “What models do you support?”, then follows up with “Which one is the fastest?” without restating the context.

One-shot answer summarization

Generate a single, concise answer to a user’s question without maintaining conversation history. This is useful when you want to display a summarized answer alongside traditional search results. Example: A user searches “How do I reset my password?” and gets a direct answer synthesized from your help articles, displayed above the regular search results.

RAG pipelines

Integrate Meilisearch as the retrieval layer in a broader RAG architecture. Meilisearch handles query understanding and hybrid retrieval, while your application controls the generation step. Example: A product recommendation engine that retrieves matching products via Meilisearch, then uses a custom prompt to generate personalized suggestions.

How it works

  1. Query understanding: Meilisearch automatically transforms the user’s natural language question into optimized search parameters
  2. Hybrid retrieval: combines keyword and semantic search for better relevancy
  3. Answer generation: your chosen LLM generates a response using only the retrieved documents as context
  4. Source attribution: every response can include references to the source documents used to generate the answer

Implementation strategies

In the majority of cases, you should use the /chats route to build conversational search. This API consolidates the entire RAG pipeline into a single endpoint, handling retrieval, context management, and generation. Follow the getting started guide to set up conversational search, then build a chat interface or generate summarized answers.

Model Context Protocol (MCP)

An alternative method is using a Model Context Protocol (MCP) server. MCPs are designed for broader uses that go beyond answering questions, but can be useful in contexts where having up-to-date data is more important than comprehensive answers. Follow the dedicated MCP guide if you want to implement it in your application.