by panabee on 2/24/24, 5:40 PM with 4 comments
by RoyTyrell on 2/24/24, 6:37 PM
by panabee on 2/24/24, 7:09 PM
Does #RAG/web search solve #LLM hallucinations?
We find that even with RAG, 45% of responses by #GPT4 to medical queries are not fully supported by retrieved URLs. The problem is much worse for GPT-4 w/o RAG, #Gemini and #Claude arxiv.org/pdf/2402.02008…
RAG ≠ faithful to source
by chrisjj on 2/24/24, 5:59 PM
Not fully supported != hallucination.