Using FastEmbed with Qdrant for Vector Search

Install Qdrant Client

pip install qdrant-client

Install FastEmbed

Installing FastEmbed will let you quickly turn data to vectors, so that Qdrant can search over them.

pip install fastembed

Initialize the client

Qdrant Client has a simple in-memory mode that lets you try semantic search locally.

from qdrant_client import QdrantClient

client = QdrantClient(":memory:")  # Qdrant is running from RAM.

Add data

Now you can add two sample documents, their associated metadata, and a point id for each.

docs = ["Qdrant has a LangChain integration for chatbots.", "Qdrant has a LlamaIndex integration for agents."]
metadata = [
    {"source": "langchain-docs"},
    {"source": "llamaindex-docs"},
]
ids = [42, 2]

Load data to a collection

Create a test collection and upsert your two documents to it.

client.add(
    collection_name="test_collection",
    documents=docs,
    metadata=metadata,
    ids=ids
)

Here, you will ask a dummy question that will allow you to retrieve a semantically relevant result.

search_result = client.query(
    collection_name="test_collection",
    query_text="Which integration is best for agents?"
)
print(search_result)

The semantic search engine will retrieve the most similar result in order of relevance. In this case, the second statement about LlamaIndex is more relevant.

[QueryResponse(id=2, embedding=None, sparse_embedding=None, 
metadata={'document': 'Qdrant has a LlamaIndex integration for agents',
'source': 'llamaindex-docs'}, document='Qdrant has a LlamaIndex integration for agents.', 
score=0.8749180370667156), 
QueryResponse(id=42, embedding=None, sparse_embedding=None, 
metadata={'document': 'Qdrant has a LangChain integration for chatbots.', 
'source': 'langchain-docs'}, document='Qdrant has a LangChain integration for chatbots.', 
score=0.8351846822959111)]
Was this page useful?

Thank you for your feedback! 🙏

We are sorry to hear that. 😔 You can edit this page on GitHub, or create a GitHub issue.