Completing the RAG Loop: From Vector Search to LLM Recommendations - Part 2

The LLM acts as the final reasoning layer, turning vector based search results into tailored viewing recommendations. Where We Left Off In Part 1, we built the foundation of a movie recommendation RAG pipeline. We took raw user preferences, generated 1024-dimension vector embeddings using bge-large, and stored them in a Qdrant vector database. The result was a semantic search engine that could find similar preferences with accuracy relevant to the user’s interests and mood of the moment. ...

September 6, 2025 · 15 min · 3076 words · Mark Holton

AI Integration: Building Conversational Apps with MCP

Why This Matters: The Shift to Agentic AI The software landscape is shifting toward agentic AI systems - applications where Large Language Models (LLMs) don’t just answer questions, but actively use tools to solve complex problems. Instead of building separate AI features, the market and industry are moving toward AI that can directly interact with your existing systems, databases, and workflows. ...

June 22, 2025 · 14 min · 2853 words · Mark Holton