[{"data":1,"prerenderedAt":380},["ShallowReactive",2],{"$fgukOamtKU1RtUiMFsqdObttmqPPQz0uc7bl_gj_LyX0":3,"$fkFxU4r5Vx76A959LbuPdXiRLW0cckk_Tl0y8SYRZobE":245,"article-222":379},{"code":4,"msg":5,"data":6},0,"",{"category":7,"tag":11,"popular":19,"latest":86,"banner":126,"list":151,"cache":244},[8,9,10],"Agent","OpenAI","LLM",[8,12,13,14,9,10,15,16,17,18],"Google","Nvidia","Claude","DeepSeek","OCR","Chat","Generator",[20,29,37,45,54,62,70,79],{"id":21,"publish_date":22,"is_original":23,"collection":5,"cover_url":24,"cover_url_1_1":25,"title":26,"summary":27,"author":28},411,"2023-09-10",1,"article_res/cover/451ef50c225a8dc61c4336506794d13b.jpeg","article_res/cover/3ba9dc7a72f87d40b20fc2d225289ee3.jpeg","Idealism","Reality is created by the mind, we can change our reality by changing our mind. - Plato","Renee's Entrepreneurial Journey",{"id":30,"publish_date":31,"is_original":23,"collection":32,"cover_url":33,"cover_url_1_1":34,"title":35,"summary":36,"author":28},108,"2024-12-07","#LLM #AGI #AI Agent","article_res/cover/0039044422e4ec9f61c18e8ee1693bb0.jpeg","article_res/cover/4220971b108a91d21407d87bb02fbaa6.jpeg","Freysa.ai: The World's First Adversarial AI Agent Game","说服 Freysa 把钱包里的钱都拿出来",{"id":38,"publish_date":39,"is_original":23,"collection":40,"cover_url":41,"cover_url_1_1":42,"title":43,"summary":44,"author":28},12,"2025-03-09","#Oxford #Reasoning #LLM #Tool Use","article_res/cover/d448e9b3617a0b5302e1bd10c438bca9.jpeg","article_res/cover/864a468f9cc4c9317efadb3811909888.jpeg","Agentic Reasoning Framework - Significantly enhance the reasoning ability of LLMs through the integration of external tools using agents","Agentic Reasoning: Reasoning LLMs with Tools for Deep Research",{"id":46,"publish_date":47,"is_original":4,"collection":48,"cover_url":49,"cover_url_1_1":50,"title":51,"summary":52,"author":53},480,"2023-04-14","#Stable Diffusion","article_res/cover/0bdbe7cb1de4a78e54536e5d9afa7ec9.jpeg","article_res/cover/b3d6ffec0608dcfaf18c5a69906d1490.jpeg","【AIGC Learning】Generate Prompts Using Word Graphs - Stable Diffusion Web UI Series 13","AI will become a powerful tool in education, transforming the way we learn and deliver instruction.  \n- Reid Hoffman","--",{"id":55,"publish_date":56,"is_original":4,"collection":57,"cover_url":58,"cover_url_1_1":59,"title":60,"summary":61,"author":28},413,"2023-09-08","#Neuroscience","article_res/cover/74f8302d78a23d9430f22171eae136b6.jpeg","article_res/cover/87ca08af81bb304746be5261160964c0.jpeg","Can machines be conscious?","Do we have an ethical obligation to not turn off conscious machines? Would turning them off be murder? No. I don't lose any sleep over unplugging a conscious machine.\n- Jeff Hawkins, \"A Thousand Brains\"",{"id":63,"publish_date":64,"is_original":23,"collection":65,"cover_url":66,"cover_url_1_1":67,"title":68,"summary":69,"author":28},178,"2024-09-09","#Entrepreneurship","article_res/cover/a7224f025b55d1820408085faef63079.jpeg","article_res/cover/11a9995b096cbf64465ef01b8673b154.jpeg","37signals company","This damn sense of relaxation",{"id":71,"publish_date":72,"is_original":4,"collection":73,"cover_url":74,"cover_url_1_1":75,"title":76,"summary":77,"author":78},460,"2023-05-12","#Google","article_res/cover/b970687b12faa52da976f91248c2aa7b.jpeg","article_res/cover/d1e71b52cfd2c63bc6e71f3e85ff135c.jpeg","Learn what BRC-20 and Ordinals are using Google Bard","Ordinals - a new protocol that allows users to store arbitrary data on the Bitcoin blockchain","Google Bard mainly writes",{"id":80,"publish_date":81,"is_original":23,"collection":5,"cover_url":82,"cover_url_1_1":83,"title":84,"summary":85,"author":28},309,"2024-03-26","article_res/cover/9877f95894ee88532d0e6012c23a2df3.jpeg","article_res/cover/20092164ddc109ce6ae56b1984246751.jpeg","Learning the Cancun Upgrade with lepton and perplexity","Building a quick conversation-based search demo with Lepton AI.",[87,95,103,111,119],{"id":88,"publish_date":89,"is_original":23,"collection":90,"cover_url":91,"cover_url_1_1":92,"title":93,"summary":94,"author":28},627,"2025-03-20","#AI Avatar #AI Video Generation","article_res/cover/d95481358f73924989f8c4ee9c75d1c8.jpeg","article_res/cover/b74bc0fab01f8b6a6aa87696c0c3ed8b.jpeg","DisPose: Generating Animated Videos by Driving Video with Reference Images","DisPose is a controllable human image animation method that enhances video generation.",{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},626,"2025-03-21","#Deep Dive into LLMs #LLM #RL #Andrej Karpathy #AlphaGo","article_res/cover/446553a5c8f8f2f07d97b20eaee84e56.jpeg","article_res/cover/e6c2823409c9b34624064b9acbaca6f1.jpeg","AlphaGo and the Power of Reinforcement Learning - Andrej Karpathy's Deep Dive on LLMs (Part 9)","Simply learning from humans will never surpass human capabilities.",{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},625,"2025-03-22","#Deep Dive into LLMs #LLM #RL #RLHF #Andrej Karpathy","article_res/cover/8da81d38b1e5cf558a164710fd8a5389.jpeg","article_res/cover/96f028d76c362a99a0dd56389e8f7a9b.jpeg","Reinforcement Learning from Human Feedback (RLHF) - Andrej Karpathy's Deep Dive on LLMs (Part 10)","Fine-Tuning Language Models from Human Preferences",{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},624,"2025-03-23","#Deep Dive into LLMs #LLM #Andrej Karpathy #AI Agent #MMM","article_res/cover/a5e7c3d48bb09109684d6513287c661d.jpeg","article_res/cover/d3f22b7c0ab8d82fd2da457a299e0773.jpeg","The Future of Large Language Models - Andrej Karpathy's In-Depth Explanation of LLM (Part 11)","preview of things to come",{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},623,"#Google #Voe #AI Video Generation","article_res/cover/c44062fea0f336c2b96b3928292392c2.jpeg","article_res/cover/a041041c69092ad3db191c5bf3ff981b.jpeg","Trial of Google's video generation model VOE2","Our state-of-the-art video generation model",[127,135,143],{"id":128,"publish_date":129,"is_original":23,"collection":130,"cover_url":131,"cover_url_1_1":132,"title":133,"summary":134,"author":28},300,"2024-04-16","#AI in Science #AGI","article_res/cover/6bf01e793e0f33e848572412eebdf9b0.jpeg","article_res/cover/91a5ee21dafecb914fabeb9430d46ec1.jpeg","Would Einstein lose his job - AI and Quantum Computing: A Glimpse into the Near Future","So Einstein's job is still safe.",{"id":136,"publish_date":137,"is_original":23,"collection":138,"cover_url":139,"cover_url_1_1":140,"title":141,"summary":142,"author":28},101,"2024-12-14","#Nvidia #AI 3D Generator","article_res/cover/693e07c85980c5c0c8fde3f037733f23.jpeg","article_res/cover/9ea8edff2d5d303ff3fffff3f6f9c3d9.jpeg","NVIDIA's open-source 3D project LLaMA-Mesh","LLaMA-Mesh: Unifying 3D Mesh Generation with Language Models",{"id":144,"publish_date":145,"is_original":23,"collection":146,"cover_url":147,"cover_url_1_1":148,"title":149,"summary":150,"author":28},131,"2024-11-10","#OpenAI","article_res/cover/87f8ed353ce39f31960e7cdfaf075a35.jpeg","article_res/cover/f597a63935f5cd32e484b4aadd6019e8.jpeg","ChatGPT has launched the Search function","Get fast, timely answers with links to relevant web sources.",{"big":152,"small":214},[153,181],{"title":154,"list":155},"AGENT",[156,157,165,173],{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":158,"publish_date":159,"is_original":23,"collection":160,"cover_url":161,"cover_url_1_1":162,"title":163,"summary":164,"author":28},622,"2025-03-24","#OWL #AI Agent #MAS #MCP #CUA","article_res/cover/cb50ca7f2bf4d1ed50202d7406e1c19a.jpeg","article_res/cover/4aa7aa3badfacf3cc84121334f1050dd.jpeg","OWL: Multi-agent collaboration","OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation",{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},620,"2025-03-26","#LLM #Google #Gemini #AI Agent","article_res/cover/53751a6dbbe990b1eb0b63f3b062aed4.jpeg","article_res/cover/031344981f0a212ff82d1f3a64aa5756.jpeg","Gemini 2.5 Pro, claimed to be far ahead of the competition, has been released with great fanfare: comprehensively surpassing other LLMs and topping the global rankings","Gemini 2.5: Our most intelligent AI model",{"id":174,"publish_date":175,"is_original":23,"collection":176,"cover_url":177,"cover_url_1_1":178,"title":179,"summary":180,"author":28},616,"2025-03-29","#MAS #AI Agent #AI Coder #MetaGPT #MGX","article_res/cover/9dcd702ad2035902e5e77967c34a1f1e.jpeg","article_res/cover/0a97fc4a922753c8f46ff38792020df8.jpeg","MGX - An automated website-building platform composed of multiple AI Agents","Your 24/7 AI Team | Dream, Chat, Create.",{"title":182,"list":183},"OPENAI",[184,191,199,206],{"id":185,"publish_date":167,"is_original":23,"collection":186,"cover_url":187,"cover_url_1_1":188,"title":189,"summary":190,"author":28},619,"#OpenAI #AI Image Generator #4o #MMM #AR Transformer","article_res/cover/2faffc97fcecf3151552cb0fd3206d89.jpeg","article_res/cover/1133cb4948af44cee2e7fbe79efb69e5.jpeg","The native image function of GPT-4o is officially launched","Introducing 4o Image Generation",{"id":192,"publish_date":193,"is_original":4,"collection":194,"cover_url":195,"cover_url_1_1":196,"title":197,"summary":198,"author":28},434,"2023-07-15","#Anthropic #OpenAI #Google #AI Code Generator #Claude","article_res/cover/e1b6f600a2b9f262a4392684e5f2ce25.jpeg","article_res/cover/6e1772e83f78f9a351ab23d3e414adee.jpeg","Latest Updates on Google Bard /Anthropic Claude2 / ChatGPT Code Interpreter","We want our models to use their programming skills to provide more natural interfaces to the basic functions of our computers.  \n - OpenAI",{"id":200,"publish_date":201,"is_original":4,"collection":146,"cover_url":202,"cover_url_1_1":203,"title":204,"summary":205,"author":28},417,"2023-08-24","article_res/cover/bccf897d50a88b18364e35f7466387e0.jpeg","article_res/cover/2f871085c1073717c1703ae86e18056f.jpeg","The GPT-3.5 Turbo fine-tuning (fine-tuning function) has been released～","Developers can now bring their own data to customize GPT-3.5 Turbo for their use cases.",{"id":207,"publish_date":208,"is_original":4,"collection":209,"cover_url":210,"cover_url_1_1":211,"title":212,"summary":213,"author":28},407,"2023-09-22","#OpenAI #AI Image Generator","article_res/cover/c59005e903d35cfc32346e2756e2728a.jpeg","article_res/cover/ba011d265e6d84b5c8cb6fd6b757b6cc.jpeg","Dall-E 3","DALL·E 3 understands significantly more nuance and detail, allowing you to easily translate your ideas into images.",[215,221,241],{"title":10,"list":216},[217,218,219,220],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":223},"GOOGLE",[224,225,226,234],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},615,"2025-03-30","#AI Researcher #AI Science #HKU #Google #AI Agent","article_res/cover/21fadf906067714bb0db31ae13a77c15.jpeg","article_res/cover/2697999a72bd26b22e85f0e92936d3ed.jpeg","AI-Researcher: LLM-driven全自动 scientific research assistant","AI-Researcher: Fully-Automated Scientific Discovery with LLM Agents  \nOpen-Sourced Alternative to Google AI Co-Scientist",{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},463,"2023-05-09","article_res/cover/89800f207723acdb55fc53bf999ebdc9.jpeg","article_res/cover/5764f369b4accd8f83e94aa4c077a175.jpeg","The Smallville sandbox world - A town with 25 virtual residents","Believable proxies of human behavior can empower interactive apps: Immersive environment, Rehearsal space, Prototyping tool",{"title":242,"list":243},"NVIDIA",[],true,{"code":4,"msg":5,"data":246},{"id":247,"publish_date":248,"is_original":23,"collection":249,"articles_id":250,"cover_url":251,"cover_url_1_1":252,"title":253,"summary":254,"author":28,"content":255,"popular":256,"list":322,"category":377,"tag":378},222,"2024-07-10","#AI Avatar #Lip Sync #Tencent","5TbVJHqH7EUvozfZWewx0w","article_res/cover/d06d42c3261157730ffe4d634dbbae8a.jpeg","article_res/cover/4c134a6c9d3f5a35c0bda53ad9397af8.jpeg","Tencent MimicMotion: High-Quality Human Motion Video Generation Based on Confidence Pose Guidance","MimicMotion : High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance","\u003Cdiv class=\"rich_media_content js_underline_content\n                       autoTypeSetting24psection\n            \" id=\"js_content\">\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>MimicMotion 能够生成任意长度的高质量视频，并模仿特定的运动引导。类似的项目之前介绍过很多：\u003C/span>\u003Cbr>\u003C/section>\u003Cul class=\"list-paddingleft-2\" style=\"list-style-type: disc;\">\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247489376&amp;idx=1&amp;sn=de5ed06770962462a49374dedfe8bf91&amp;chksm=c13d13baf64a9aacd0e562774326f435e7abff51595d3156de987ce747289506f33574e8a476&amp;scene=21#wechat_redirect\" textvalue=\"快手的LivePortrait - 视频驱动Avatar动画框架\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">快手的LivePortrait - 视频驱动Avatar动画框架\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247489178&amp;idx=1&amp;sn=66b6faf1ec9959452ad145c9278f31dc&amp;chksm=c13d1240f64a9b56a7ece4b65bd0fb0d1015539d7937fe0a6faaf3cb659ee71b29e7ccb0679b&amp;scene=21#wechat_redirect\" textvalue=\"MusePose和Follow-Your-Pose：腾讯发布的姿势驱动人物动作\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">MusePose和Follow-Your-Pose：腾讯发布的姿势驱动人物动作\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247488927&amp;idx=1&amp;sn=a706f31471bd0c8bbe5fb06ee46b2edb&amp;chksm=c13d1145f64a985326ef70b246c0ce3e311f82c9c841e0bf5bd4fccbdd18e230fcbafc2b0641&amp;scene=21#wechat_redirect\" textvalue=\"让 Avatar 动起来 - InstructAvatar、EMO、Follow-Your-Emoji\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">让 Avatar 动起来 - InstructAvatar、EMO、Follow-Your-Emoji\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247488926&amp;idx=1&amp;sn=b8c34b1660b9cd15d2a3c77993fd7ebf&amp;chksm=c13d1144f64a9852aaaae90c4729e267289fcdecaa1962731ceb4185fc234d7acc17c6291fcb&amp;scene=21#wechat_redirect\" textvalue=\"腾讯的两个让 Avatar 说话的模型：V-Express 和 MuseTalk\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">腾讯的两个让 Avatar 说话的模型：V-Express 和 MuseTalk\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247488924&amp;idx=1&amp;sn=7ca9212c451ab6fd3b6e6d126ee344e5&amp;chksm=c13d1146f64a985095d4305386d6cdac8950e34de5db6e4a9cab1b9e06b67af7b17f90199ba6&amp;scene=21#wechat_redirect\" textvalue=\"VASA-1 微软的逼真音频驱动的实时生成会话面孔\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">VASA-1 微软的逼真音频驱动的实时生成会话面孔\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247488209&amp;idx=1&amp;sn=7553b6dcbe5ff825da1e11c6ff36c6b3&amp;chksm=c13d160bf64a9f1d76d59d3abc60ef4ec7b0c0433a5a0220f7ecf4700689cf1e17dd933418eb&amp;scene=21#wechat_redirect\" textvalue=\"【Synthesia最新功能】表情丰富的虚拟人和语音克隆\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">【Synthesia最新功能】表情丰富的虚拟人和语音克隆\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247487736&amp;idx=1&amp;sn=12b99bfcf33982c06a9a356292d42f23&amp;chksm=c13d1422f64a9d34f4bf1de9bd647a3fb3e939fa8926f5c124b9b7163c79b34b5390f632e157&amp;scene=21#wechat_redirect\" textvalue=\"Rhubarb Lip Sync - AI 生成二次元角色的口型动画\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">Rhubarb Lip Sync - AI 生成二次元角色的口型动画\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247487709&amp;idx=1&amp;sn=4aa174705d31e6195d56ef5ae83c71c9&amp;chksm=c13d1407f64a9d1101f80d0d51da4146760856037d042587f12118847063a085cf33bb59c07a&amp;scene=21#wechat_redirect\" textvalue=\"Champ - 人体图像静态图片生成动画\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">Champ - 人体图像静态图片生成动画\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247487690&amp;idx=1&amp;sn=b33cedc3278337d11a7743ccd3fe0852&amp;chksm=c13d1410f64a9d06f7e3fed6e18363d216086b3bfe423a30f41f7a6962b9db00a2b3dd2cde3d&amp;scene=21#wechat_redirect\" textvalue=\"AniPortrait - 音频驱动的真实感肖像动画合成技术\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">AniPortrait - 音频驱动的真实感肖像动画合成技术\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247487027&amp;idx=1&amp;sn=85940f34a5834497022b325ec4fc3099&amp;chksm=c13d0ae9f64a83ffd42318268dcef530594042a5db14b47a304a09b07436fa6ed6e5466db571&amp;scene=21#wechat_redirect\" textvalue=\"阿里的DreaMoving：基于扩散模型的人物视频生成框架\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">阿里的DreaMoving：基于扩散模型的人物视频生成框架\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247486983&amp;idx=1&amp;sn=d059317d76ec29b1e1a838b0fadd72eb&amp;chksm=c13d0addf64a83cb48886eb11781b395c019f2498bb1347580126b4995473bffbe65b5c6e7cc&amp;scene=21#wechat_redirect\" textvalue=\"Meta 的 Audio2Photoreal - 从声音到虚拟人动起来\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">Meta 的 Audio2Photoreal - 从声音到虚拟人动起来\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247486443&amp;idx=1&amp;sn=a6f1ca9fe096793f0e4062d5ba81ec25&amp;chksm=c13d0f31f64a8627553543c628d1c60fc319cb723dcc45a9f031ae08d2047bc18fa31fc10c64&amp;scene=21#wechat_redirect\" textvalue=\"微软 GAIA：ZERO-SHOT的对口型单张图生成视频技术\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">微软 GAIA：ZERO-SHOT的对口型单张图生成视频技术\u003C/a>\u003C/span>\u003C/section>\u003C/li>\u003Cli>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247486443&amp;idx=1&amp;sn=a6f1ca9fe096793f0e4062d5ba81ec25&amp;chksm=c13d0f31f64a8627553543c628d1c60fc319cb723dcc45a9f031ae08d2047bc18fa31fc10c64&amp;scene=21#wechat_redirect\" textvalue=\"微软 GAIA：ZERO-SHOT的对口型单张图生成视频技术\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">\u003C/a>\u003Ca target=\"_blank\" href=\"http://mp.weixin.qq.com/s?__biz=MzkwOTMzMzk0MQ==&amp;mid=2247485313&amp;idx=1&amp;sn=b36e78cef530b058c5e79ba82e0462a3&amp;chksm=c13d035bf64a8a4d1081372e4f7cab990b6b3f4b9e1ddc71ffb5e017498c825d905b61571aa6&amp;scene=21#wechat_redirect\" textvalue=\"使用 SadTalker 生成数字人视频\" linktype=\"text\" imgurl=\"\" imgdata=\"null\" data-itemshowtype=\"0\" tab=\"innerlink\" data-linktype=\"2\">使用 SadTalker 生成数字人视频\u003C/a>\u003C/span>\u003C/section>\u003Csection>\u003Cspan style='background-color: rgb(255, 255, 255);color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;'>\u003Cbr>\u003C/span>\u003C/section>\u003C/li>\u003C/ul>\u003Ch2 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">本项目亮点：\u003C/span>\u003C/h2>\u003Cul style='margin-top: 8px;margin-bottom: 8px;cursor: pointer;padding-left: 25px;color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);' class=\"list-paddingleft-1\">\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">引入了置信度姿态引导，确保了高帧质量和时间连续性。\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">引入了基于姿态置信度的区域损失放大，大大减少了图像失真。\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">为了生成长且平滑的视频，提出了一种渐进式潜在融合策略。\u003C/section>\u003C/li>\u003C/ul>\u003Ch2 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">Showcases\u003C/span>\u003C/h2>\u003Cul style='margin-top: 8px;margin-bottom: 8px;cursor: pointer;padding-left: 25px;color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);' class=\"list-paddingleft-1\">\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">动作\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803228620.09890526070297567.mp4\" poster=\"./assets/17423803228830.26105572006881506.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803228730.2561944745030338.mp4\" poster=\"./assets/17423803228620.36108448711940966.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">跳舞\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803231580.5338827647271693.mp4\" poster=\"./assets/17423803228610.1999275985634532.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803247280.6404522826052332.mp4\" poster=\"./assets/17423803229490.942550891098054.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">说话\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803254640.07475771312154644.mp4\" poster=\"./assets/17423803231960.06797127821691307.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17423803276170.24665159729769526.mp4\" poster=\"./assets/17423803240430.5885379386022176.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003C/li>\u003C/ul>\u003Ch2 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003C/h2>\u003Ch2 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">试用\u003C/span>\u003C/h2>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>可以在 Replicate - https://replicate.com/zsxkib/mimic-motion 上运行 MimicMotion 进行试用。\u003C/p>\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img\" data-galleryid=\"\" data-imgfileid=\"100005811\" data-ratio=\"1.5450180072028812\" data-s=\"300,640\" data-type=\"png\" data-w=\"833\" style=\"\" src=\"./assets/17423803277870.612760515634504.png\">\u003C/p>\u003Cp>\u003Cbr>\u003C/p>\u003Cp>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">方法\u003C/span>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005796\" data-ratio=\"0.5296296296296297\" data-type=\"png\" data-w=\"1080\" src=\"./assets/17423803277870.37717322098916206.png\">\u003C/span>\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cstrong style=\"cursor: pointer;background: none 0% 0% / auto no-repeat scroll padding-box border-box rgba(0, 0, 0, 0);width: auto;height: auto;border-style: none;border-width: 3px;border-color: rgba(0, 0, 0, 0.4);border-radius: 0px;\">MimicMotion\u003C/strong> 结合了图像到视频的扩散模型和新颖的置信度姿态引导。模型的可训练组件包括一个时空 U-Net 和一个用于引入姿态序列作为条件的 PoseNet。置信度姿态引导的关键特性包括：\u003C/p>\u003Col style='margin-top: 8px;margin-bottom: 8px;cursor: pointer;padding-left: 25px;color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);' class=\"list-paddingleft-1\">\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">姿态序列伴随着关键点置信度评分，使模型能够根据评分自适应地调整姿态引导的影响。\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">高置信度的区域在损失函数中赋予更大的权重，增强其在训练中的影响。\u003C/section>\u003C/li>\u003C/ol>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">置信度姿态引导\u003C/span>\u003C/h3>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>我们利用姿态引导帧的亮度来表示姿态估计的置信度。\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005797\" data-ratio=\"0.6916666666666667\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803277890.6739429611849039.jpeg\">\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>这种设计增强了生成对错误引导信号的鲁棒性（姿态 1&amp;2），并提供了可见性提示以解决姿态模糊问题（姿态 3）。\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005804\" data-ratio=\"0.6796296296296296\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803280480.47282352836589014.jpeg\">\u003C/p>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">特定区域的手部细化\u003C/span>\u003C/h3>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>我们实施了一种基于置信度阈值生成掩码的策略。对于置信度评分超过预设阈值的区域，我们取消掩码，从而识别出可靠的区域。在计算视频扩散模型的损失时，相应于未遮蔽区域的损失值会被放大，使其在模型训练中比其他遮蔽区域更有效。\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>在相同的参考图像和姿态引导下，进行手部增强训练能够显著减少手部失真并提升视觉效果。\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005805\" data-ratio=\"0.32407407407407407\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803278730.5874999640941121.jpeg\">\u003C/p>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">渐进式潜在融合以实现时间平滑\u003C/span>\u003C/h3>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>我们提出了一种渐进式的方法来生成具有时间平滑效果的长视频。在每个去噪步骤中，视频片段首先在训练模型的条件下分别去噪，参考相同的图像和相应的姿态子序列。在每个去噪步骤中，图中用虚线框标记的重叠帧根据它们的位置逐步融合。\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005798\" data-ratio=\"0.36203703703703705\" data-type=\"png\" data-w=\"1080\" src=\"./assets/17423803278750.6216005818600536.png\">\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;text-wrap: wrap;background-color: rgb(255, 255, 255);'>渐进式潜在融合实现了平滑过渡，避免了视频片段边界的突然变化，从而增强了长视频生成的整体视觉时间一致性。\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005806\" data-ratio=\"0.5462962962962963\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803278740.5796953285876665.jpeg\">\u003C/p>\u003Ch2 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 22px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">对比\u003C/span>\u003C/h2>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">定性评估\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005799\" data-ratio=\"0.3527777777777778\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803290730.6813441687982833.jpeg\">\u003Cimg class=\"rich_pages wxw-img\" data-imgfileid=\"100005800\" data-ratio=\"0.5777777777777777\" data-type=\"jpeg\" data-w=\"1080\" src=\"./assets/17423803290830.36772871113213657.jpeg\">\u003C/span>\u003C/h3>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;text-wrap: wrap;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">定量评估\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img\" data-galleryid=\"\" data-imgfileid=\"100005802\" data-ratio=\"0.425\" data-s=\"300,640\" data-type=\"png\" data-w=\"800\" style=\"\" src=\"./assets/17423803281710.9463144945598374.png\">\u003C/p>\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img\" data-galleryid=\"\" data-imgfileid=\"100005803\" data-ratio=\"0.3679060665362035\" data-s=\"300,640\" data-type=\"png\" data-w=\"1022\" style=\"\" src=\"./assets/17423803278770.8658113288148901.png\">\u003C/p>\u003C/span>\u003C/h3>\u003Cp style=\"display: none;\">\u003Cmp-style-type data-value=\"3\">\u003C/mp-style-type>\u003C/p>\u003C/div>",[257,266,274,281,289,298,306,314],{"id":258,"title_md5":259,"publish_date":260,"author_md5":261,"is_original":23,"collection":262,"summary_md5":263,"cover_url":264,"cover_url_1_1":265},84,"c1c7ee59ace8f8a90289685e4d5ee0a1","2024-12-31","bc27fa490c4d0d525bac812fc0793534","#Robotics #Embodied AI #Nvidia","a792c2c2621d23d2dbd779ecbaaaa612","article_res/cover/9674435f69d7a9096f49e354619b1b4b.jpeg","article_res/cover/792b21b2cfc9af32151ee3d0c4a85864.jpeg",{"id":267,"title_md5":268,"publish_date":269,"author_md5":261,"is_original":23,"collection":270,"summary_md5":271,"cover_url":272,"cover_url_1_1":273},454,"91641111b0c792baa0b33ffe39da2b8c","2023-05-23","#AI Audio Generator","e88bd758169331024d1e41e1d2ee3871","article_res/cover/1e70875d4849ffa4f5d888fb70dfa02f.jpeg","article_res/cover/b8cdd05f2b0e85ca053d0dedba60f57f.jpeg",{"id":275,"title_md5":276,"publish_date":277,"author_md5":261,"is_original":4,"collection":5,"summary_md5":278,"cover_url":279,"cover_url_1_1":280},440,"2625737f4c4b641898eb12f03ddd434f","2023-06-24","3514c1dce56b91ffba23126867b67fef","article_res/cover/e5ac56e1640c601a518a9fb24370ec3e.jpeg","article_res/cover/659be0c95bfea0624423348bdd82434c.jpeg",{"id":282,"title_md5":283,"publish_date":284,"author_md5":261,"is_original":23,"collection":285,"summary_md5":286,"cover_url":287,"cover_url_1_1":288},64,"3fc45b522a322e3997ff98bc49a6c595","2025-01-21","#AGI #Stuart Russell","ca5a1e91d9b50b0b713a47dfc4421581","article_res/cover/77971cb66a8e920986f6795aa11ce6f6.jpeg","article_res/cover/f8264767787e1a53e5d8eaf95e2dbec7.jpeg",{"id":290,"title_md5":291,"publish_date":292,"author_md5":293,"is_original":4,"collection":294,"summary_md5":295,"cover_url":296,"cover_url_1_1":297},484,"af76b06a6a201755883f5b64a78911eb","2023-04-10","cfab1ba8c67c7c838db98d666f02a132","#Stable Diffusion #AI Image Generator","5d615395f499310d37a158d7addf51f7","article_res/cover/73624a211e1da591990a68332c54be3e.jpeg","article_res/cover/8d10285c2263513b4ec9b69fcc93a2c1.jpeg",{"id":299,"title_md5":300,"publish_date":301,"author_md5":261,"is_original":23,"collection":302,"summary_md5":303,"cover_url":304,"cover_url_1_1":305},36,"7ac2688dcb2412a28f10570ab0fc0ca2","2025-02-17","#Baidu #AI Avatar","bfc395b4cdb996740934b1e9328f30c9","article_res/cover/2f0a10000836de9daa5902903690697c.jpeg","article_res/cover/19bcf2b18f917a8e7e375cb69a00687d.jpeg",{"id":307,"title_md5":308,"publish_date":309,"author_md5":261,"is_original":23,"collection":310,"summary_md5":311,"cover_url":312,"cover_url_1_1":313},39,"3f58518f52d1307800bfc7df809b8998","2025-02-14","#Deep Dive into LLMs #Andrej Karpathy #LLM #Tool Use #Hallucination","919a69ba5f8bc9300ce8f8ab21a7ddee","article_res/cover/11e858ad6b74dfa80f923d549b62855c.jpeg","article_res/cover/615e1b320f1fc163edc1d2d154a6de33.jpeg",{"id":315,"title_md5":316,"publish_date":317,"author_md5":261,"is_original":23,"collection":318,"summary_md5":319,"cover_url":320,"cover_url_1_1":321},25,"78ee9c62989c96701d84c5911874010c","2025-02-27","#Alibaba #AI Video Editor #Animate Anyone #AI Video Generation","9cfbe107bd45f72e2967e6a339e18f8f","article_res/cover/450c28bb351bd7dd1aa9ab908b494204.jpeg","article_res/cover/6bfa652b1433ba6033e45ac848042745.jpeg",{"related":323,"small":362},[324,331,340,347,354],{"id":325,"publish_date":326,"is_original":23,"collection":270,"cover_url":327,"cover_url_1_1":328,"title":329,"summary":330,"author":28},453,"2023-05-24","article_res/cover/9f9d88dde4a788562db422dbf0a2872d.jpeg","article_res/cover/29ac820986dde6a7bff6e359574cef18.jpeg","Train Your Own Voice: SoftVC VITS Singing Voice Conversion Fork","A fork of so-vits-svc with real-time support and a greatly improved interface.",{"id":332,"publish_date":333,"is_original":4,"collection":334,"cover_url":335,"cover_url_1_1":336,"title":337,"summary":338,"author":339},476,"2023-04-18","#AI 3D Generator","article_res/cover/4bf87096649f9f42597ca3bec98185f1.jpeg","article_res/cover/596ca8d93d90fe4f40ee9eefd72a222b.jpeg","Various AIGC 3D tools (Part 1)","GPT creates realistic 3D models and renders of buildings based on sketches and specifications.  \n-","AIGC Learning Notes",{"id":341,"publish_date":342,"is_original":4,"collection":5,"cover_url":343,"cover_url_1_1":344,"title":345,"summary":346,"author":28},378,"2023-11-27","article_res/cover/a6e956c7192e834c23b9791f362717b6.jpeg","article_res/cover/fb444ef951eb25829e83c372cb080eb2.jpeg","Create GPTs using Zapier","Equip GPTs to take action in 3rd party apps with Zapier AI Actions!",{"id":348,"publish_date":349,"is_original":23,"collection":146,"cover_url":350,"cover_url_1_1":351,"title":352,"summary":353,"author":28},383,"2023-11-07","article_res/cover/b980267b1641c318fd033ba204c1ad22.jpeg","article_res/cover/9200a2f7d1cf6eef1e06d0bfcf8f76e4.jpeg","Key Updates from OpenAI DevDay","New platform updates from DevDay: GPT-4 Turbo, Assistants API, new modalities + GPTs in ChatGPT",{"id":355,"publish_date":356,"is_original":4,"collection":5,"cover_url":357,"cover_url_1_1":358,"title":359,"summary":360,"author":361},575,"2022-04-11","article_res/cover/cfa59ca34acfab4ce41315c5be30d2d1.jpeg","article_res/cover/63be7726eb61f99334648faf767dd8cb.jpeg","Translation and Summary of the Messari Report 【Options 1-8 for Entering the Mainstream Market】","Will coins outperform the companies that support them?","Translation",[363,369,375],{"title":10,"list":364},[365,366,367,368],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":370},[371,372,373,374],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},{"title":242,"list":376},[],[8,9,10],[8,12,13,14,9,10,15,16,17,18],["Reactive",245],1754646409428]