[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tag-video-generation":3},{"tag":4,"articles":11},{"id":5,"name":6,"slug":7,"article_count":8,"description_zh":9,"description_en":10},"9d786a9c-c6f0-4c37-9f0a-2aab262eabce","video generation","video-generation",4,"影片生成正在從「會動」走向「可控、可解釋」：研究重點包括時間一致性、速度與節奏控制、相機與物件運動分離，以及主動／被動動作建模，讓生成結果更接近真實世界的因果關係與使用者意圖。","Video generation is shifting from making clips that merely move to models that can be controlled and reasoned about. Current work focuses on temporal consistency, playback speed, camera-object motion separation, and active vs. passive actions to better match user intent and real-world causality.",[12,21,29],{"id":13,"slug":14,"title":15,"summary":16,"category":17,"image_url":18,"cover_image":18,"language":19,"created_at":20},"be28a180-07a1-433c-bb6a-6f015d7291c2","actcam-joint-camera-motion-control-zh","ActCam 讓鏡頭和動作一起控","ActCam 用零樣本方式，同時控制影片生成中的角色動作與攝影機路徑，重點是不必重新訓練新模型。","research","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778220675487-z4sz.png","zh","2026-05-08T06:10:33.672754+00:00",{"id":22,"slug":23,"title":24,"summary":25,"category":26,"image_url":27,"cover_image":27,"language":19,"created_at":28},"c8b244e2-00b3-44f1-b418-53eda4350cb4","how-to-migrate-from-sora-2-in-2026-zh","2026 如何遷移 Sora 2","把 Sora 2 的影片工作流遷移到新模型，並在 OpenAI 停用期限前完成備份、測試與替換。","tools","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778137244364-kchy.png","2026-05-07T07:00:29.44463+00:00",{"id":30,"slug":31,"title":32,"summary":33,"category":17,"image_url":34,"cover_image":34,"language":19,"created_at":35},"ff7256d7-69ae-40a3-abf5-c2e7a487e4c7","moright-motion-control-causality-zh","MoRight 讓影片控制更有因果感","MoRight 把相機運動和物件運動拆開，還把動作分成主動與被動兩段，讓影片生成不只會動，還能更像真的有因果關係。","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775714993444-7ipn.png","2026-04-09T06:09:31.667703+00:00"]