Wednesday, March 26, 2025 at 12:00:00 AM
Lately, I've been toying with this portfolio, which is supposed to be AI-powered. Sort of.
In the world of AI and machine learning, Retrieval-Augmented Generation (RAG) promises to bridge the gap between raw language models and contextually rich, accurate responses. My recent experience revealed a somewhat critical flaw: not all embeddings are created equal. What with my visualization past experience, I thought "who knows! Maybe visualization can be the key to understanding what is wrong?"
Like many developers exploring RAG, I started with off-the-shelf solutions. I was using Ollama and Mistral models, expecting seamless performance. The reality? Disappointingly generic responses that missed the nuanced context of my specific document collections. Hallucinations? You bet!
Frustrated by mediocre results, I began developing a custom embedding visualization tool. The goal is simple yet theoretically powerful: understand how my embeddings actually represent my documents.
At its core, an embedding is a dense vector representation of text. Each document becomes a point in a high-dimensional space, where semantic similarities create clusters and relationships. I recommend you read about it, the internet is full of interesting content.
I created a what I call a "wave visualization" that represents embedding dimensions on a horizontal axis. Each point is an aggregate of a document's embedding and is put on the Y axis as different vector dimensions. The visualization reveals:
The visualization uses:
// Simplified cosine similarity calculation
const calculateCosineSimilarity = (vec1: number[], vec2: number[]) => {
const dotProduct = vec1.reduce((sum, val, i) => sum + val * vec2[i], 0);
const magnitude1 = Math.sqrt(vec1.reduce((sum, val) => sum + val * val, 0));
const magnitude2 = Math.sqrt(vec2.reduce((sum, val) => sum + val * val, 0));
return dotProduct / (magnitude1 * magnitude2);
};
My current approach is a work in progress. Next steps include:
Embedding visualization isn't just a debugging tool—it's a window into the semantic understanding of my AI system. By seeing how documents are represented, I can (theoretically) make more informed decisions about our RAG pipelines.
Stay tuned for more updates on this embedding exploration journey!