[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-hallucinations":3},{"tag":4,"articles":9},{"id":5,"name":6,"slug":6,"article_count":7,"description_zh":8,"description_en":8},"488b4a27-bb00-4e30-9266-07bd56efae53","hallucinations",0,null,[10],{"id":11,"slug":12,"title":13,"summary":14,"category":15,"image_url":16,"cover_image":16,"language":17,"created_at":18},"fcba2ffc-9687-40b6-b58c-a36dc8b4926b","retrieval-augmented-generation-explained-en","Retrieval-Augmented Generation, Explained Simply","RAG lets large language models pull fresh facts from documents before answering, which cuts hallucinations and adds citations.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778083860476-4o28.png","en","2026-05-06T16:10:34.177377+00:00"]