PinnedBest Practices for RAG PipelineOver the past few years, RAG has matured and multiple studies has been done to understand pattern and behaviors which can result in low…Sep 1, 2024Sep 1, 2024
PinnedHow Agentic RAG solves problem with current RAG limitationsIn this volume 4 of coffee break concept, we will understand how AgenticRAG helps solve limitations of traditional RAG.Aug 17, 20241Aug 17, 20241
PinnedHow Much GPU Memory is Needed to Serve a Large Language Model (LLM)?In nearly all LLM interviews, there’s one question that consistently comes up: “How much GPU memory is needed to serve a Large Language…Aug 17, 202418Aug 17, 202418
Will Long-Context LLMs Make RAG Obsolete?Long-Context LLMs — models capable of processing context windows up to 1 million tokens — poses an intriguing question: Will Long-Context…Nov 19, 2024Nov 19, 2024
11 Chunking Strategies for RAG — Simplified & VisualizedRetrieval-Augmented Generation (RAG) combines pre-trained language models with information retrieval systems to produce more accurate and…Nov 2, 2024Nov 2, 2024
Mastering Caching Methods in Large Language Models (LLMs)Large Language Models (LLMs) like OpenAI’s GPT-4 have transformed natural language processing, enabling applications ranging from chatbots…Sep 27, 2024Sep 27, 2024
How to Select Right LLM model for your use caseWhen you begin any client project, one of the most frequently asked questions is, “Which model should I use?” There isn’t a straightforward…Sep 7, 2024Sep 7, 2024
How OpenAI or DeepMind calculates cost of training a transformer based models?The basic equation giving the cost to train a transformer model is given by:Aug 24, 2024Aug 24, 2024
New course on AgenticRAG with LlamaIndexA new course on agenticRAG with LlamaIndex includes 5 real-time case studies with code examplesAug 15, 2024Aug 15, 2024
Tired of Poor RAG Results?If you are tired of poor RAG results then follow these steps, a coffee break concepts Vol 2Jun 9, 2024Jun 9, 2024