To ensure the system retrieves relevant results, Lucidworks recommends you implement comprehensive testing of semantic vector searches.
ImportantThis feature is currently only available to clients who have contracted with Lucidworks for features related to Neural Hybrid Search and Lucidworks AI.
This feature is only available in Managed Fusion 5.9.x for versions 5.9.6+.
Is used as a benchmark to determine vector search performance
Represents frequent queries and variations to ensure optimal relevancy
Contains ground truth that defines the most relevant results for each query
Most effective when it contains a significant number of diverse queries that are updated periodically to reflect the most pertinent information
Enhanced when automated testing frameworks are incorporated, which can result in more extensive coverage of search scenarios and provide continuous system performance monitoring
The dataset needs to include a wide range of queries that reflect your organization’s real-world user interactions. The types of queries to include are:
Typical use cases that comprise frequent queries
A variety of queries that test system rules and functions such as misspellings, ambiguous terms, synonyms, and phrasing differences
Query results from vector searches using the golden dataset are compared to the ground truth data. The system calculates metrics that provide information to help you enhance the search and return more relevant results.Typically, performance metrics can be categorized as follows:
Precision metrics focus on results that adhere most closely with the criteria specified in the query. The most relevant items are reported in the top results. For example, precision metrics display the top 3 results.
Recall metrics evaluate the overall results retrieved and how relevant they are to ground truth specifications.
Ranking analysis metrics report results in order of relevance, with the most relevant results ranked the highest.
Lucidworks Platform clients can also view metrics for custom trained models.While these metrics are about the associated model, and not the golden dataset, they provide a solid basis for understanding if your model is suitable for deployment with your dataset.