[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-gemini-25-flash":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"07af57e2-d09a-4f8a-abb2-a8572e638690","Gemini 2.5 Flash","gemini-25-flash",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"3cb0da95-801d-485d-9583-539027365723","why-ai-safety-teams-are-wrong-blame-only-alignment-en","Why AI safety teams are wrong to blame only alignment","AI models do not just fail from bad alignment; they also inherit harmful stories from training data.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778947422376-naaj.png","en","2026-05-16T16:03:17.251356+00:00"]