[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-hidden-representations":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"f66153d5-dab2-4f98-b327-0c4463b02472","hidden representations","hidden-representations",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"22c43f4e-8be9-4440-bd1b-74a00b60dfa3","llms-implicit-grammar-representations-en","Do LLMs Learn Grammar Beyond Likelihood?","A probe study finds hidden layers in language models encode grammaticality better than string probability, but not plausibility.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778135464967-fzem.png","en","2026-05-07T06:30:35.804749+00:00"]