[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-linear-probes":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"344bf532-a39c-46c9-a6c2-3566bbb75f0e","linear probes","linear-probes",0,null,[11],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"22c43f4e-8be9-4440-bd1b-74a00b60dfa3","llms-implicit-grammar-representations-en","Do LLMs Learn Grammar Beyond Likelihood?","A probe study finds hidden layers in language models encode grammaticality better than string probability, but not plausibility.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778135464967-fzem.png","en","2026-05-07T06:30:35.804749+00:00"]