Member-only story

Fully Local RAG with Qdrant, Ollama, LangChain and LangServe

Agent Issue
9 min readNov 10, 2023

--

xAI has just unveiled Grok, LLM inspired by Hitchhiker’s Guide to the Galaxy!

Grok’s unique feature is its ability to access real-time knowledge, a capability powered by Qdrant — open-source vector similarity search engine and vector database written in Rust.

Join our next cohort: Full-stack GenAI SaaS Product in 4 weeks!

You can think of Qdrant as a specialized tool for storing, searching, and managing vectors, designed to be both fast and reliable, even under substantial workloads.

You can couple Qdrant with LangChain and Ollama to build applications that can leverage LLMs and semantic search, which you can deploy locally or in the cloud — pretty powerful stuff!

In this article, I’ll guide you through several key processes:

  1. Handling PDF documents
  2. Creating and storing embeddings with GPT4AllEmbeddings and storing them with Qdrant
  3. Serving LLMs such as Llama2 7b Chat locally with Ollama
  4. Setting up a local Retrieval Augmented Generation (RAG) application using LangChain and LangServe.

Ready to dive? Let’s begin!

Getting Started with LangChain, Ollama and Qdrant

--

--

Agent Issue
Agent Issue

Written by Agent Issue

Your front-row seat to the future of Agents.

No responses yet