Memgraph logo
Back to blog
RAG isn’t dead. It’s just getting started.

RAG isn’t dead. It’s just getting started.

By Dominik Tomicevic
7 min readApril 16, 2025

I say LLMs are amazing. They can crank out plausible answers, write better code comments than your junior dev, and even draft blog posts like this one (don’t worry, this one’s all me). But there’s a massive blind spot no one’s talking about enough: LLMs don’t know your data.

No matter how advanced OpenAI’s next model or the latest open-source darling like DeepSeek gets, these systems are still trained on general, public data. And that’s fine if you’re asking about the capital of Peru or who won the Super Bowl in 2008. But when you’re asking questions about your company, your workflows, or your proprietary systems - they’ve got nothing. Worse, they make up answers that sound convincing but are completely wrong.

And this isn’t some fringe issue, it’s a systemic flaw. LLMs can’t tell you what they don’t know, and they can’t reason their way out of an incorrect answer. That’s especially a big problem if you’re an enterprise relying on them for real-world, high-stakes decisions.

So, here’s how I think about it:

If we’re serious about making AI useful in the real world, we need to stop chasing magic and start focusing on context. And context isn’t magic—it’s structure.

LLMs Alone Aren’t Enough

The real risk with LLMs isn’t that they occasionally get things wrong. It’s that they sound confidently right when they’re wrong. And that’s dangerous. Imagine this:

You ask the LLM: “Is this transaction suspicious?” It replies, “Yes, because it resembles known patterns.” Sounds legit, right? But did it analyze relationships between accounts, historical behavior, or hidden loops? No. Did it even have access to that data to be able to answer that? No. At least not entirely. It’s just regurgitating patterns from training data.

Or here’s a better example from a pharmaceutical research domain. You ask the LLM to summarize clinical trial results and predict drug efficacy and you get “This combination has shown a 30% improvement.” Except those trials were never run together, and critical side effects were overlooked. That’s not just misinformation; it’s malpractice. You see my point?

Let me explain further.

LLMs Are The Clever Parrots of AI

Imagine you’re teaching a parrot to solve puzzles. The parrot can repeat what you’ve shown it, even in creative ways, but it doesn’t truly understand the puzzle—it’s just mimicking patterns. That’s how LLMs work. They’re very clever parrots, capable of producing answers that look like reasoning, but they don’t genuinely understand or reason the way humans do.

Some of the “clever tricks” LLMs use to seem smarter include:

  • Thinking out loud (chain-of-thought prompting): It’s like watching a parrot say to itself, “First, I’ll do this. Then I’ll do that.” It sounds logical and can improve performance on multi-step questions, but it’s just repeating patterns it learned—it’s not coming up with the reasoning itself.
  • Using examples (few-shot learning): Show the parrot three examples of solving a puzzle, and it might copy those steps to solve the fourth. But if the puzzle changes, the parrot is stuck—it doesn’t actually understand the rules. Constantly updating the underlying model with new data and hints can give the system gradual improvement, but the parrot still isn’t "thinking."
  • Pretending to think (simulated reasoning): Some models act as if they’re reasoning by breaking their answers into steps. It’s like a parrot saying, “Hmm, let me think,” before giving a response. It looks thoughtful but is still just pattern-matching.
  • Learning from other parrots (synthetic data): Imagine one parrot teaches another what it learned from humans. The second parrot might sound smarter, but it’s just repeating what it learned secondhand—mistakes and all.
  • Fancy Wrapping (pseudo-structure): Some models format their answers in structured ways (like adding tags around steps) to appear more organized. It’s like a parrot putting its sentences in bold—neat, but not smarter.

These tricks are like sleight of hand—they make the model seem brilliant but don’t fix the core issue: the model doesn’t actually understand what it’s saying.

Why RAG Still Matters (And Always Will)

This is where RAG is important. RAG makes AI safer, more scalable, and enterprise-ready.

If you work for an enterprise and you’ve been tasked to whip up something with RAG for an enterprise setting, you’ve probably already done some digging. And you’ve probably realized a few things.

1. Security and scaling aren’t optional

I bet your enterprise deals with sensitive data—financial reports, legal contracts, proprietary workflows, whatever. Without fine-grained access control, an AI system might serve up confidential information to the wrong person. That’s a breach waiting to happen.

If you know how to implement RAG, you will set up strict access policies so only authorized users can access specific datasets. Also, that sensitive data stays private, no matter how complex the query. Without this, enterprises risk fines, lawsuits, and reputational damage. And nobody wants to explain to their board why a chatbot leaked trade secrets.

2. Hallucinations are a dealbreaker

Another risk with LLMs is that they can and most often do, generate plausible-sounding nonsense. RAG fixes this by grounding AI outputs in real, verifiable data. Instead of relying on probabilistic predictions, RAG retrieves structured, trustworthy information.

3. Thinking of fine-tuning LLMs? Don’t.

Fine-tuning LLMs on enterprise data sounds great in theory. In practice? It’s expensive, slow, and doesn’t scale. By the time you’ve retrained your model, the workflows or datasets have already changed. WithRAG, you skip the retraining nightmare. RAG dynamically retrieves only what’s relevant from a constantly updated database.

So regardless of what anybody tells you, RAG isn’t here to replace LLMs, and LLMs aren’t here to replace RAG. They work together to create a better together.

Why Graph RAG Is the Future

There, I’ve said it. And I know I’ve spent decades working with graph databases, so I might be a bit biased—but I also know I’m right. Graphs are the ultimate context machine. They’re not just databases; they’re dynamic networks of meaning, so to say. They don’t just store data; they show how everything is connected.

Instead of sifting through a mountain of data, GraphRAG finds what’s relevant to your specific question. It’s almost like having a personalized researcher pulling insights for you. Also… GraphRAG integrates real-time data changes without retraining, ensuring the context is always up-to-date.

And there you have it. In finance, Graph RAG detects fraud by analyzing relationships across accounts, not just patterns in text. In healthcare, AI needs secure access to patient data, compliance rules, and treatment protocols. With Graph RAG, doctors, admins, and compliance officers get tailored, role-specific answers without risking sensitive data. We at Memgraph have two such use cases:

Context Is King

Bigger models aren’t the full answer. Better context is. Structure and context are the key to turning generative models into something actually useful. RAG (and especially Graph RAG) is how we get there. By pairing generative AI with structured, verifiable knowledge, businesses can build AI systems that are smarter, safer, and future-proof.

LLMs, no matter how advanced, will never “just know” your internal data. They weren’t trained on it, and they never will be. RAG is what connects them to your corporate reality.

So no, RAG isn’t dead. It’s just getting started. The question isn’t whether you need it—it’s how quickly you can start using it.

Getting Started with Memgraph

Join us on Discord!
Find other developers performing graph analytics in real time with Memgraph.
© 2025 Memgraph Ltd. All rights reserved.