Avi Chawla
Search documents
X @Avi Chawla
Avi Chawla· 2026-01-25 11:51
If you found it insightful, reshare it with your network.Find me → @_avichawlaEvery day, I share tutorials and insights on DS, ML, LLMs, and RAGs. https://t.co/lgCr54DUEVAvi Chawla (@_avichawla):Vector Index vs Vector Database, clearly explained!Devs typically use these terms interchangeably.But understanding this distinction is necessary since it leads to problems down the line.Here's how to think about it:A vector index is basically a search algorithm.You https://t.co/wI5HOVlUyt ...
X @Avi Chawla
Avi Chawla· 2026-01-25 06:31
Milvus GitHub repo:(don't forget to star 🌟)https://t.co/QWS0u9aHCK ...
X @Avi Chawla
Avi Chawla· 2026-01-25 06:31
Vector Index vs Vector Database, clearly explained!Devs typically use these terms interchangeably.But understanding this distinction is necessary since it leads to problems down the line.Here's how to think about it:A vector index is basically a search algorithm.You give it vectors, it organizes them into something searchable (like HNSW), and it finds similar items fast. FAISS is another example.But here's the thing.That's all it does. It doesn't handle storage, it doesn't filter by metadata, and it doesn't ...
X @Avi Chawla
Avi Chawla· 2026-01-24 06:45
Download Voyage-4-nano from HF: https://t.co/mXjgxriy6d ...
X @Avi Chawla
Avi Chawla· 2026-01-24 06:45
100%That's how much data you re-index when you change embedding models.And most teams treat that as unavoidable.Imagine you built a RAG pipeline using an embedding model with high retrieval quality, and it ships to production.Six months later, a better embedding model is released that delivers similar quality at a lower cost.But your existing embeddings live in one vector space, while the new model produces embeddings in a different one, which makes them incompatible.Switching models now means rebuilding th ...
X @Avi Chawla
Avi Chawla· 2026-01-23 18:30
RT Avi Chawla (@_avichawla)Researchers built a new RAG approach that:- does not need a vector DB.- does not embed data.- involves no chunking.- performs no similarity search.And it hit 98.7% accuracy on a financial benchmark (SOTA).Here's the core problem with RAG that this new approach solves:Traditional RAG chunks documents, embeds them into vectors, and retrieves based on semantic similarity.But similarity ≠ relevance.When you ask "What were the debt trends in 2023?", a vector search returns chunks that ...
X @Avi Chawla
Avi Chawla· 2026-01-22 06:51
The efficacy of mixed precision scaling over traditional training is evident from the image below.Mixed precision training is over 2.5x faster than conventional training in a mini neural network.Typical speeds are 4-6x in bigger neural networks.Check this 👇 https://t.co/RRGQZ3bYLI ...
X @Avi Chawla
Avi Chawla· 2026-01-22 06:50
A simple technique trains neural nets 4-6x faster!- OpenAI used it in GPT models.- Meta used it in LLaMA models.- Google used it in Gemini models.Here's a breakdown (with code): ...