[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-gpt-41":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"2033eb79-8281-452f-b27a-3efcd10af7f7","GPT-4.1","gpt-41",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"3cb0da95-801d-485d-9583-539027365723","why-ai-safety-teams-are-wrong-blame-only-alignment-en","Why AI safety teams are wrong to blame only alignment","AI models do not just fail from bad alignment; they also inherit harmful stories from training data.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778947422376-naaj.png","en","2026-05-16T16:03:17.251356+00:00"]