Vector databases enable us to store information as embeddings and search for the “results similar” to our input query using. These results are then provided to the LLM as context so it can respond in a context-aware manner using Retrieval Augmented Generation (RAG).

Our goal is to search relevant information from a knowledge base quickly, here’s how vector databases are used with LLMs:

1

Chunk the information

Break down the knowledge into smaller chunks to ensure our search query matches only relevant results.

2

Load the knowledge base

Convert the chunks into embedding vectors and store them in a vector database.

3

Search the knowledge base

When the user sends a message, we convert the input message into an embedding and “search” for nearest neighbors in the vector database.

The following VectorDb are supported: