AI-Infused Applications: The Rise of Vector and Graph Databases


Traditional databases are fantastic when it comes to storing and retrieving structured data—think numbers, dates, or neatly organized text that matches exactly. But when we step into the world of AI, especially in fields like natural language processing (NLP) and image recognition, the game changes.

AI works with unstructured data—things like text, images, audio, and video —that don’t fit neatly into rows and columns. To make sense of this kind of data, AI needs to go beyond surface-level information and truly understand its meaning and context—what we call understanding it semantically.

💡 The term
semantically
is the key here. It refers to the understanding of data in terms of its meaning and context rather than just its literal representation. So how does a vector database capture *semantic* meanings of a given content? 🤔 The answer is
vector embeddings
.

Vector Embeddings
These are numerical representations of data capturing the semantic meaning or features of that data. For example, the vector embeddings of "dog" 🐕 and "puppy" 🐶 would be closer than the embeddings of "dog" 🐕 and "mango" 🥭. Here's an example of an vector embedding:
cat=[1.5,−0.4,7.2,19.6,3.1]

How vector embeddings are generated?
There are specialized machine learning models for the vectorization process. Eg: Word2Vec, GloVe.

💡 Usually in Retrieval Augmented Generation (RAG) applications, it doesn't make much sense to convert an entire coprus of text or individual words to vector embeddings. Instead large corpus (exceeding 42,000 character) of text get broken down into manageable chunks of text so that a similarity search can be performed among the chunks.

Searching/Querying in a vector DB
Once the text is converted into vector embeddings the resulting embeddings (text segments) will then be stored in the vector db along with their
metadata
. This process typically should happen during the ingestion phase. After ingestion we can query the db using natural language and the vector db will spit out relevant
text segments
by doing
similarity search
. The similarity search algorithm is typically based on
Cosine similarity
or
Euclidean distance
. At this point you might be wondering if the database stores the data in the form of vectors of number how can the database understand our natural language query. Well it doesn't. The database only knows about numbers. So in order the database to do a similarity search we have to convert our query to a vector embeddings too. Then only the db can do the similarity search and return similar vector embeddings.

Relevance
We can instruct the DB to only return a given max number of results and more importantly the contextual relevance. The relevance is specifed by floating point number between zero to one and higher value means highly relevant. For eg: you can say get me the top 3 results with relevance greater than 0.7

Piecing it all together

At this point we have a user query and a set of text results from a vector db similarity search ready to be fed into an LLM to get an answer. So the
retrieval
part is done. We need to
augment
the context with the results and ask the LLM to use the augmented context to answer the user query. This can be done by altering the original prompt (user query) Eg:
Answer the following user query by only using the information provided below: 

- {{ text chunks }}

User query is: {{ query }}

Knowledge graphs

At this point you can jump into creating a RAG pipeline right away and the LLM would have adequate information to answer straight factual questions.

Use cases for RAG
  • Chat bot answering questions about refund policy
  • Customer support for product information
  • Simple FAQ systems

However a mere similarity search would not yield some of the intricate details that may be crucial in providing an accurate answer.

Consider following story:

  • John was happily married to Katie for ten years
  • Katie is a childhood friend of Martha
  • Martha is Jake’s wife
  • Jake and Trevor are best friends
  • Trevor had a fight with John
  • After Jake learned about this fight, he started hating Trevor
The dislike seems to stem from Jake's loyalty to his social network. Since John is married to Katie, who is a close friend of Jake's wife Martha, Jake likely feels a sense of loyalty to John by extension. When Trevor (Jake's best friend) had a fight with John, Jake took John's side over Trevor's, causing Jake to start hating Trevor.

If this information was spread across a 200-page book that has been vectorized by paragraphs, a similarity search for “Why Jake dislikes Trevor so much” would likely return these specific lines:

  • “Jake and Trevor are best friends”
  • “Trevor had a fight with John”
  • “Jake learns about the fight and Jake started hating Trevor”
  • “Martha is Jake’s wife”

💡 Notice that above results do not reveal the intricate detail of why Jake end friendship with Trevor other than the fight. “John was happily married to Katie for ten years” and “Martha and Katies friendship goes back to childhood” Indirectly ties Katie/John to Jake’s social circle, potentially explaining why Jake cares about the fight. This is where a graph database can help.

How a Graph Database Helps

A graph database would be particularly powerful in this situation because it excels at capturing and navigating complex relationships between entities - exactly what we need to understand Jake's motivation. In this story, the reason for Jake's strong reaction isn't directly stated but depends on understanding the social connections. Let see how a graph database would potentially model this:

Nodes: Jake, Martha, Trevor, John, Katie

Edges:

  • Jake --MARRIED_TO--> Martha
  • Martha --FRIEND_OF--> Katie
  • Katie --MARRIED_TO--> John
  • Trevor --HAD_FIGHT_WITH--> John
  • Jake --FRIENDS_OF--> Trevor

This structure visualizes chains of loyalty: Jake → Martha → Katie → John vs. Jake → Trevor The fight between Trevor and John disrupts this web, forcing Jake to “choose sides.”

Building a knowledge graph

This is one of the challanging part and takes the greatest amount of time. We need a help of a
Natural Language Processing (NLP)
model to extract key entities and relationships. The output of this process can be a knowledge triplets (subject, predicate, object) or a json. Subject and object are entities and the relationships are called predicates. In addition to this, an entity can have various properties Eg: Person entity can have DOB as a property. Example extraction:
Text
: Mr.A is the CEO of company Y.
Triplet
: (Mr.A, WORKS_AT, Y)

Graph retrieval

First we need to extract entities from the given user prompt. This can be achieved by asking an LLM to extract the entities from the user query and building the query programatically. The other easier method is to ask the LLM to generate a graph query in Cypher or Gremlin and execute that in the graph db. Then these entities and their relationships can be fetched from the graph database. Remember to add a full text index on the entity name so that text search can run on the entity names. Since full text indices can match with language specific stemming and stopwords.
Here's a graph query:
{
  people_search(func: eq(entity, "John Doe")) {
    entity
    entity_type
    relation
    object
    object_type
  }
}

and an example output:

{
  "data": {
    "people_search": [
      {
        "entity": "John Doe",
        "entity_type": "Person",
        "relation": "OCCUPATION",
        "object": "Businessman",
        "object_type": "Occupation"
      },
      {
        "entity": "John Doe",
        "entity_type": "Person",
        "relation": "DEATH_CAUSE",
        "object": "Murder",
        "object_type": "CauseOfDeath"
      }
    ]
  }
}

Similar to what we do to the results of vector embeddings we can embed the query output into the final prompt so that the LLM will have much more details pertaining to the given user question. We may need to limit the depth of the query to filter out unwanted results which however will take some trial and error to when to use what depth.

Conclusion

In this guide, we’ve explored the world of vector databases and their applications in text retrieval. We’ve seen how vector embeddings can be used to search for relevant text segments, and how the results can be further augmented with contextual information using Large Language Models (LLMs).

We’ve also delved into the realm of graph databases, which excel at capturing complex relationships between entities. By modeling these relationships as nodes and edges, we can create a powerful knowledge graph that can help us answer nuanced questions.

The key takeaways from this guide are:

  • Vector databases offer a powerful way to search for relevant text segments
  • LLMs can be used to augment the context with additional information
  • Graph databases provide a powerful way to capture complex relationships between entities
  • Knowledge graphs can help us answer nuanced questions by modeling these relationships as nodes and edges

To build a comprehensive knowledge graph, we’ll need to:

  • Extract key entities and relationships from user prompts using NLP models
  • Model these relationships as nodes and edges in a graph database
  • Add full-text indices on entity names for efficient text search
  • Use LLMs to augment the context with additional information


Content under the appendix is only meant for my reference for future post. If you find my guide useful please do share it :)

Appendix


Graph databases
Qdrant - a vector database

Qdrant is a relatively new addition to vector database space. What I like the most about Qdrant is how easy it is to get started development using it. You just download the platform binary here and launch 🚀

ulimit -n 10000
./qdrant

I had to set the ulimit to avoid getting error: Too many open files 👀

Getting the frontend working

Although in official docs say that Qdrant’s web ui client can be accessed through http://localhost:6333/dashboard it doesn’t work out of the box at the time of writing. You have to clone their web gui repo and run npm:

git clone https://github.com/qdrant/qdrant-web-ui
cd qdrant-web-ui
npm i
npm start

Now you should be able to access it through http://localhost:5173 (you need to run qdrant in a separate terminal :duh 🙄 )