<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Improve Search on Qdrant - Vector Search Engine</title><link>https://qdrant.tech/documentation/improve-search/</link><description>Recent content in Improve Search on Qdrant - Vector Search Engine</description><generator>Hugo</generator><language>en-us</language><managingEditor>info@qdrant.tech (Andrey Vasnetsov)</managingEditor><webMaster>info@qdrant.tech (Andrey Vasnetsov)</webMaster><atom:link href="https://qdrant.tech/documentation/improve-search/index.xml" rel="self" type="application/rss+xml"/><item><title>Measuring Retrieval Relevance</title><link>https://qdrant.tech/documentation/improve-search/retrieval-relevance/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://qdrant.tech/documentation/improve-search/retrieval-relevance/</guid><description>&lt;h1 id="measuring-retrieval-relevance"&gt;Measuring Retrieval Relevance&lt;/h1&gt;
&lt;table&gt;
 &lt;thead&gt;
 &lt;tr&gt;
 &lt;th&gt;Time: 40 min&lt;/th&gt;
 &lt;th&gt;Level: Intermediate&lt;/th&gt;
 &lt;th&gt;&lt;/th&gt;
 &lt;th&gt;&lt;/th&gt;
 &lt;/tr&gt;
 &lt;/thead&gt;
 &lt;tbody&gt;
 &lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;This tutorial focuses on &lt;strong&gt;retrieval relevance&lt;/strong&gt;: how well retrieved results match real user intent.
To measure retrieval relevance, you need a labeled dataset of queries paired with their expected relevant documents (commonly called a &lt;em&gt;golden query set&lt;/em&gt; or &lt;em&gt;ground truth&lt;/em&gt;). This tutorial covers both building that dataset and running it through Qdrant to compute relevance metrics.&lt;/p&gt;
&lt;p&gt;Two related tutorials cover the other retrieval-evaluation concerns: &lt;a href="https://qdrant.tech/documentation/tutorials-search-engineering/ann-recall/"&gt;Measuring ANN Recall&lt;/a&gt; (does the approximate index match exact kNN?) and &lt;a href="https://qdrant.tech/documentation/improve-search/pipeline-output-quality/"&gt;Evaluating Pipeline Output Quality&lt;/a&gt; (does the end-to-end pipeline produce the right output?).&lt;/p&gt;</description></item><item><title>Evaluating Pipeline Output Quality</title><link>https://qdrant.tech/documentation/improve-search/pipeline-output-quality/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://qdrant.tech/documentation/improve-search/pipeline-output-quality/</guid><description>&lt;h1 id="evaluating-pipeline-output-quality"&gt;Evaluating Pipeline Output Quality&lt;/h1&gt;
&lt;table&gt;
 &lt;thead&gt;
 &lt;tr&gt;
 &lt;th&gt;Time: 45 min&lt;/th&gt;
 &lt;th&gt;Level: Intermediate&lt;/th&gt;
 &lt;th&gt;&lt;/th&gt;
 &lt;th&gt;&lt;/th&gt;
 &lt;/tr&gt;
 &lt;/thead&gt;
 &lt;tbody&gt;
 &lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;This tutorial focuses on &lt;strong&gt;pipeline output quality&lt;/strong&gt;: whether the full retrieval pipeline produces the right output once retrieved results reach a consumer, most often an LLM generator in a RAG system.
To measure pipeline output quality, you run your golden set through the full pipeline, capture each &lt;code&gt;(question, retrieved_context, answer)&lt;/code&gt; triple, and score the triples against judgment metrics like faithfulness, answer relevancy, and context precision.&lt;/p&gt;</description></item></channel></rss>