That’s a great post, thanks for breaking this down.
I’ve seen similar hallucination issues pop up when the RAG pipeline doesn’t enforce proper context control or document-level isolation. If you're building anything in a sensitive or enterprise context, it might be worth looking into tools that provide stronger safeguards around that.
I work at a company called TeraVera, and we just launched a secure AI platform with a zero-trust design and tenant-level data segregation. It was built specifically to prevent things like model hallucinations and unauthorized data blending in RAG applications—especially helpful in regulated industries.
If you're interested, here’s a link to request API access and check out the dev docs: teravera.com/api-access-form/
Main site: teravera.com
Hope that helps!