Back to home

Tag

RAG

RAG, or retrieval-augmented generation, combines search with model output so answers can be grounded in source data. This tag covers vector databases, hybrid search, indexing, recall, and evaluation for knowledge bases, semantic search, and enterprise assistants.

22 articles

How to Add Temporal RAG in Production
AI Agent/May 13

How to Add Temporal RAG in Production

Add a temporal reranking layer to RAG so fresh, valid, and versioned facts rank correctly.

RAGFlow adds agents to open-source RAG
Tools & Apps/May 12

RAGFlow adds agents to open-source RAG

RAGFlow pairs retrieval-augmented generation with agent features, Docker self-hosting, and support for newer models like GPT-5 and Gemini 3 Pro.

Why RAG in Microsoft Foundry needs better indexes, not bigger prompts
Industry News/May 9

Why RAG in Microsoft Foundry needs better indexes, not bigger prompts

RAG in Microsoft Foundry succeeds when retrieval is indexed well, not when prompts get longer.

Why ChatGPT’s Goblin Bug Proves Closed Models Are Fragile
Industry News/May 9

Why ChatGPT’s Goblin Bug Proves Closed Models Are Fragile

ChatGPT’s goblin bug shows why closed LLMs are too opaque for serious production use.

How to Build Advanced RAG in n8n
AI Agent/May 8

How to Build Advanced RAG in n8n

Build a production RAG pipeline in n8n with chunking, hybrid retrieval, reranking, and compression.

Why RAG is ending for agentic AI
AI Agent/May 7

Why RAG is ending for agentic AI

RAG is the wrong layer for agentic AI, and compilation-stage knowledge systems will replace it.

Why RAG Needs a Self-Healing Layer, Not Just Better Prompts
Research/May 7

Why RAG Needs a Self-Healing Layer, Not Just Better Prompts

RAG should be treated as a failure-prone system that needs real-time self-healing, not prompt tuning.

Why Open-Source LLMs Must Be Judged by Workload, Not Hype
Research/May 7

Why Open-Source LLMs Must Be Judged by Workload, Not Hype

Open-source LLMs in 2026 should be chosen by workload fit, not benchmark hype.

Retrieval-Augmented Generation, Explained Simply
Research/May 7

Retrieval-Augmented Generation, Explained Simply

RAG lets large language models pull fresh facts from documents before answering, which cuts hallucinations and adds citations.

RAG precision tuning can hurt retrieval accuracy
Research/May 6

RAG precision tuning can hurt retrieval accuracy

Redis research says tuning RAG embeddings for precision can cut retrieval accuracy by up to 40% and weaken agentic pipelines.

AWS Bedrock Knowledge Bases simplifies RAG
Tools & Apps/May 5

AWS Bedrock Knowledge Bases simplifies RAG

Amazon Bedrock Knowledge Bases helps teams build RAG apps with managed ingestion, retrieval, citations, and structured-data queries.

Why Databricks RAG Is a Platform Play, Not a Feature
Industry News/May 5

Why Databricks RAG Is a Platform Play, Not a Feature

Databricks treats RAG as an end-to-end platform problem, and that is the right way to build it.

How to Build a RAG Pipeline in 5 Steps
AI Agent/May 5

How to Build a RAG Pipeline in 5 Steps

Build a retrieval-augmented generation pipeline that grounds AI answers in your own data.

Actian’s VectorAI DB Claims 22x Faster Search
Tools & Apps/May 5

Actian’s VectorAI DB Claims 22x Faster Search

Actian says VectorAI DB embeds vector search inside apps and delivers up to 22x faster retrieval than common alternatives.

Why the 2026 AI engineer roadmap is the wrong starting point
Industry News/May 4

Why the 2026 AI engineer roadmap is the wrong starting point

The 2026 AI engineer roadmap is too broad to be the first plan you follow.

MathNet: Global Multimodal Math Reasoning & Retrieval
Research/Apr 21

MathNet: Global Multimodal Math Reasoning & Retrieval

MathNet adds 30,676 Olympiad problems across 47 countries and tests both solving and retrieval for multimodal models.

Qdrant vs Milvus vs Weaviate for RAG in 2026
Tools & Apps/Apr 14

Qdrant vs Milvus vs Weaviate for RAG in 2026

Qdrant, Milvus, and Weaviate power different RAG needs in 2026. Here’s how they compare on latency, scale, hybrid search, and cost.

IBM hits 100B vectors on one server
Research/Apr 14

IBM hits 100B vectors on one server

IBM says its CAS prototype indexed 100 billion vectors on one server, with 694 ms latency and 90% recall for RAG.

How Windsurf Flow Keeps Context Alive
Tools & Apps/Apr 4

How Windsurf Flow Keeps Context Alive

Windsurf Flow updates AI context as you work. Here’s how RAG, Cascade, Memories, and rules shape every suggestion.

What OpenRAG Does for Enterprise AI
Tools & Apps/Apr 3

What OpenRAG Does for Enterprise AI

IBM’s OpenRAG packages retrieval, indexing, and model orchestration so teams can build grounded AI apps on their own data.

20 GitHub AI Projects to Watch in 2026
Tools & Apps/Mar 26

20 GitHub AI Projects to Watch in 2026

OpenClaw may top GitHub, but 2026’s AI list shows a bigger shift toward agents, workflow systems, RAG, and multimodal tools.

RAG in 2026: The Indispensable AI Bridge
Model Releases/Mar 26

RAG in 2026: The Indispensable AI Bridge

In 2026, advanced Retrieval Augmented Generation (RAG) systems are essential for bridging large language models with enterprise knowledge, ensuring informed AI outputs.