[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-epistemic-guardrails":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"e4e73beb-7ab9-4c0c-ae72-d03f2c0492a6","epistemic guardrails","epistemic-guardrails",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"19e8a080-c448-4e41-92f5-9ec578bed83b","ai-reading-assistants-epistemic-guardrails-en","Why AI reading assistants need guardrails","A minimal prototype tests whether LLM reading assistants stay honest when users push them beyond retrieval into interpretation.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778050257881-qv24.png","en","2026-05-06T06:50:41.442428+00:00"]