[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-lora":3},{"tag":4,"articles":10},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":9},"20157975-ca90-488d-aa19-f6fa6467cb61","LoRA","lora",2,null,[11,20],{"id":12,"slug":13,"title":14,"summary":15,"category":16,"image_url":17,"cover_image":17,"language":18,"created_at":19},"bfbcb15a-47ab-478e-822a-38d89dc8cb84","lora-vs-qlora-vs-full-fine-tuning-zh","LoRA vs QLoRA vs 全量微調","這篇比較 LoRA、QLoRA 與全量微調，幫你用成本、顯存、速度與效果判斷哪一種大語言模型微調方式最適合你的團隊。","industry","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778915627798-evv7.png","zh","2026-05-16T07:13:32.474543+00:00",{"id":21,"slug":22,"title":23,"summary":24,"category":25,"image_url":26,"cover_image":26,"language":18,"created_at":27},"868034d7-415b-49bd-8f25-4dbd602e7094","unsloth-qwen35-partial-fine-tuning-zh","Unsloth 讓 Qwen3.5 可分層微調","Unsloth 新增 Qwen3.5 視覺模型分層微調，能只訓練 vision、language、attention 或 MLP。VRAM 更省，訓練也更快，對多模態團隊很實用。","tools","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775218014686-wj6q.png","2026-04-03T12:06:38.523525+00:00"]