This quickstart guide walks you through setting up Lucidworks AI using Neural Hybrid Search (NHS) and retrieval-augmented generation (RAG). You’ll prepare your data, vectorize documents and queries, configure query blending, and enable AI-powered responses. Each step builds a working configuration that combines lexical relevance with semantic understanding for search, and generates natural language answers using your own content. In just a few steps, you’ll have a functional pipeline you can test, tune, and use as a starting point for future implementations.
1

Prepare your documents

As your documents are indexed, they need to be prepared for Lucidworks AI. This includes and .Both steps run in one stage called the LWAI Chunker Stage. Add it to your index pipeline anywhere before the Solr Partial Update Indexer stage.

LWAI Chunker Stage stage in the correct position in the index pipeline.

Configuration is straightforward, but if you need additional guidance, expand the related sections in this step.Select the chunking strategy that fits your use case. Lucidworks recommends starting with the Sentence strategy. You can try other strategies as needed.Choose a model for vectorization. You don’t need the most advanced model to get excellent results. Start with snowflake_l_chunk_vector_1024v and adjust as needed. Make sure the model fits your use case and goals.Set the Input context variable field to the location where Fusion should store the vectors. For example, <doc.embedding_t>.Index some documents and continue to the next step.
2

Vectorize your queries

Just as you needed to vectorize documents, you need to vectorize queries so Lucidworks AI can find the best matching document chunks.Add the LWAI Vectorize Query stage before any stage that alters the query in your query pipeline. Use the same model you selected for the LWAI Chunker Stage in the index pipeline.

LWAI Vectorize Query stage in the correct position in the query pipeline.

Now, your queries are vectorized for Neural Hybrid Search (NHS) and retrieval-augmented generation (RAG).
3

Configure NHS

Neural Hybrid Search combines and queries to create a flexible balance for any use case.Add the Chunking Neural Hybrid Query stage anywhere between the LWAI Vectorize Query and Solr Query stages. The default configuration is a good starting point, but you must set the Vector Query Field to match the value in the Destination Field Name & Context Output field from the LWAI Chunker Stage.

Chunking Neural Hybrid Query stage in the correct position in the query pipeline.

Enter several queries and test the results. Then adjust the lexical and vector and as needed. Keep experimenting until you find the right balance for your use case.
4

Configure RAG

Add the LWAI Prediction stage anywhere between the LWAI Vectorize Query and Solr Query stages. For this use case, select RAG.

LWAI Prediction stage in the correct position in the query pipeline.

Choose a model that fits your retrieval-augmented generation needs. It does not have to match the model used for indexing. Start with a model that aligns with your goals, and switch to more advanced options if needed. For additional guidance, expand the related section in this step.
Lucidworks AI supports many use cases. To explore more, see Lucidworks AI use cases.
Lucidworks AI needs specific information to perform RAG. In the Document Field Mappings section, configure at least the body and source fields. You can include additional fields if they help improve the responses.If your model requires an API key, add it to the configuration. The key is stored securely.
5

Fine-tune

Everything is set up. Now test your configuration in the Query Workbench. Switch to the JSON view to inspect Neural Hybrid Search results and RAG responses.

RAG responses in the Query Workbench.

Enter queries and evaluate the responses. Neural Hybrid Search should return relevant results. If the results are weak, adjust the query weights and squash factors.Lucidworks AI is flexible and supports many use cases. Use the following checks to guide your evaluation:
  1. Is the answer backed by retrieved content? The response should only include facts found in the retrieved documents.
  2. Do citations match the content? References must point to documents that support the answer.
  3. What happens if nothing useful is retrieved? The system should avoid generating unsupported content.
  4. Does the answer stay on topic? The response should directly address the query.
  5. Does the system handle edge cases well? Use ambiguous or off-topic queries to test its behavior.
If results are still unsatisfactory, adjust the model configuration. If issues persist, review your model selection.