[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-data-quality":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"54a71871-b0d7-4a12-aa73-5b6146e41761","data quality","data-quality",0,null,[11,20],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"d3d5812b-849a-4a6e-8c8c-d859618bd4b2","why-fine-tuning-llms-domain-tasks-right-default-en","Why fine-tuning LLMs for domain tasks is the right default","Fine-tuning is the best default when an LLM must be accurate in a narrow domain.","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778916227001-iu04.png","en","2026-05-16T07:23:33.047894+00:00",{"id":21,"slug":22,"title":23,"summary":24,"category":16,"image_url":25,"cover_image":25,"language":18,"created_at":26},"18fb2e62-3d41-4b4c-8d65-e91e5f20ea28","microsoft-goalcover-fine-tuning-gaps-en","Microsoft’s GoalCover finds fine-tuning gaps","Microsoft Research’s GoalCover spots missing capabilities in fine-tuning data before training, and improved Qwen-3-14B reward scores.","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778462450292-viev.png","2026-05-11T01:20:34.483926+00:00"]