[{"data":1,"prerenderedAt":374},["ShallowReactive",2],{"$fgukOamtKU1RtUiMFsqdObttmqPPQz0uc7bl_gj_LyX0":3,"$fPr2FTEvKeJNoEO-gFC8wdE1GwXgaCkk3qSa5zCunU3M":245,"article-627":373},{"code":4,"msg":5,"data":6},0,"",{"category":7,"tag":11,"popular":19,"latest":86,"banner":126,"list":151,"cache":244},[8,9,10],"Agent","OpenAI","LLM",[8,12,13,14,9,10,15,16,17,18],"Google","Nvidia","Claude","DeepSeek","OCR","Chat","Generator",[20,29,37,45,54,62,70,79],{"id":21,"publish_date":22,"is_original":23,"collection":5,"cover_url":24,"cover_url_1_1":25,"title":26,"summary":27,"author":28},411,"2023-09-10",1,"article_res/cover/451ef50c225a8dc61c4336506794d13b.jpeg","article_res/cover/3ba9dc7a72f87d40b20fc2d225289ee3.jpeg","Idealism","Reality is created by the mind, we can change our reality by changing our mind. - Plato","Renee's Entrepreneurial Journey",{"id":30,"publish_date":31,"is_original":23,"collection":32,"cover_url":33,"cover_url_1_1":34,"title":35,"summary":36,"author":28},108,"2024-12-07","#LLM #AGI #AI Agent","article_res/cover/0039044422e4ec9f61c18e8ee1693bb0.jpeg","article_res/cover/4220971b108a91d21407d87bb02fbaa6.jpeg","Freysa.ai: The World's First Adversarial AI Agent Game","说服 Freysa 把钱包里的钱都拿出来",{"id":38,"publish_date":39,"is_original":23,"collection":40,"cover_url":41,"cover_url_1_1":42,"title":43,"summary":44,"author":28},12,"2025-03-09","#Oxford #Reasoning #LLM #Tool Use","article_res/cover/d448e9b3617a0b5302e1bd10c438bca9.jpeg","article_res/cover/864a468f9cc4c9317efadb3811909888.jpeg","Agentic Reasoning Framework - Significantly enhance the reasoning ability of LLMs through the integration of external tools using agents","Agentic Reasoning: Reasoning LLMs with Tools for Deep Research",{"id":46,"publish_date":47,"is_original":4,"collection":48,"cover_url":49,"cover_url_1_1":50,"title":51,"summary":52,"author":53},480,"2023-04-14","#Stable Diffusion","article_res/cover/0bdbe7cb1de4a78e54536e5d9afa7ec9.jpeg","article_res/cover/b3d6ffec0608dcfaf18c5a69906d1490.jpeg","【AIGC Learning】Generate Prompts Using Word Graphs - Stable Diffusion Web UI Series 13","AI will become a powerful tool in education, transforming the way we learn and deliver instruction.  \n- Reid Hoffman","--",{"id":55,"publish_date":56,"is_original":4,"collection":57,"cover_url":58,"cover_url_1_1":59,"title":60,"summary":61,"author":28},413,"2023-09-08","#Neuroscience","article_res/cover/74f8302d78a23d9430f22171eae136b6.jpeg","article_res/cover/87ca08af81bb304746be5261160964c0.jpeg","Can machines be conscious?","Do we have an ethical obligation to not turn off conscious machines? Would turning them off be murder? No. I don't lose any sleep over unplugging a conscious machine.\n- Jeff Hawkins, \"A Thousand Brains\"",{"id":63,"publish_date":64,"is_original":23,"collection":65,"cover_url":66,"cover_url_1_1":67,"title":68,"summary":69,"author":28},178,"2024-09-09","#Entrepreneurship","article_res/cover/a7224f025b55d1820408085faef63079.jpeg","article_res/cover/11a9995b096cbf64465ef01b8673b154.jpeg","37signals company","This damn sense of relaxation",{"id":71,"publish_date":72,"is_original":4,"collection":73,"cover_url":74,"cover_url_1_1":75,"title":76,"summary":77,"author":78},460,"2023-05-12","#Google","article_res/cover/b970687b12faa52da976f91248c2aa7b.jpeg","article_res/cover/d1e71b52cfd2c63bc6e71f3e85ff135c.jpeg","Learn what BRC-20 and Ordinals are using Google Bard","Ordinals - a new protocol that allows users to store arbitrary data on the Bitcoin blockchain","Google Bard mainly writes",{"id":80,"publish_date":81,"is_original":23,"collection":5,"cover_url":82,"cover_url_1_1":83,"title":84,"summary":85,"author":28},309,"2024-03-26","article_res/cover/9877f95894ee88532d0e6012c23a2df3.jpeg","article_res/cover/20092164ddc109ce6ae56b1984246751.jpeg","Learning the Cancun Upgrade with lepton and perplexity","Building a quick conversation-based search demo with Lepton AI.",[87,95,103,111,119],{"id":88,"publish_date":89,"is_original":23,"collection":90,"cover_url":91,"cover_url_1_1":92,"title":93,"summary":94,"author":28},627,"2025-03-20","#AI Avatar #AI Video Generation","article_res/cover/d95481358f73924989f8c4ee9c75d1c8.jpeg","article_res/cover/b74bc0fab01f8b6a6aa87696c0c3ed8b.jpeg","DisPose: Generating Animated Videos by Driving Video with Reference Images","DisPose is a controllable human image animation method that enhances video generation.",{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},626,"2025-03-21","#Deep Dive into LLMs #LLM #RL #Andrej Karpathy #AlphaGo","article_res/cover/446553a5c8f8f2f07d97b20eaee84e56.jpeg","article_res/cover/e6c2823409c9b34624064b9acbaca6f1.jpeg","AlphaGo and the Power of Reinforcement Learning - Andrej Karpathy's Deep Dive on LLMs (Part 9)","Simply learning from humans will never surpass human capabilities.",{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},625,"2025-03-22","#Deep Dive into LLMs #LLM #RL #RLHF #Andrej Karpathy","article_res/cover/8da81d38b1e5cf558a164710fd8a5389.jpeg","article_res/cover/96f028d76c362a99a0dd56389e8f7a9b.jpeg","Reinforcement Learning from Human Feedback (RLHF) - Andrej Karpathy's Deep Dive on LLMs (Part 10)","Fine-Tuning Language Models from Human Preferences",{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},624,"2025-03-23","#Deep Dive into LLMs #LLM #Andrej Karpathy #AI Agent #MMM","article_res/cover/a5e7c3d48bb09109684d6513287c661d.jpeg","article_res/cover/d3f22b7c0ab8d82fd2da457a299e0773.jpeg","The Future of Large Language Models - Andrej Karpathy's In-Depth Explanation of LLM (Part 11)","preview of things to come",{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},623,"#Google #Voe #AI Video Generation","article_res/cover/c44062fea0f336c2b96b3928292392c2.jpeg","article_res/cover/a041041c69092ad3db191c5bf3ff981b.jpeg","Trial of Google's video generation model VOE2","Our state-of-the-art video generation model",[127,135,143],{"id":128,"publish_date":129,"is_original":23,"collection":130,"cover_url":131,"cover_url_1_1":132,"title":133,"summary":134,"author":28},300,"2024-04-16","#AI in Science #AGI","article_res/cover/6bf01e793e0f33e848572412eebdf9b0.jpeg","article_res/cover/91a5ee21dafecb914fabeb9430d46ec1.jpeg","Would Einstein lose his job - AI and Quantum Computing: A Glimpse into the Near Future","So Einstein's job is still safe.",{"id":136,"publish_date":137,"is_original":23,"collection":138,"cover_url":139,"cover_url_1_1":140,"title":141,"summary":142,"author":28},101,"2024-12-14","#Nvidia #AI 3D Generator","article_res/cover/693e07c85980c5c0c8fde3f037733f23.jpeg","article_res/cover/9ea8edff2d5d303ff3fffff3f6f9c3d9.jpeg","NVIDIA's open-source 3D project LLaMA-Mesh","LLaMA-Mesh: Unifying 3D Mesh Generation with Language Models",{"id":144,"publish_date":145,"is_original":23,"collection":146,"cover_url":147,"cover_url_1_1":148,"title":149,"summary":150,"author":28},131,"2024-11-10","#OpenAI","article_res/cover/87f8ed353ce39f31960e7cdfaf075a35.jpeg","article_res/cover/f597a63935f5cd32e484b4aadd6019e8.jpeg","ChatGPT has launched the Search function","Get fast, timely answers with links to relevant web sources.",{"big":152,"small":214},[153,181],{"title":154,"list":155},"AGENT",[156,157,165,173],{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":158,"publish_date":159,"is_original":23,"collection":160,"cover_url":161,"cover_url_1_1":162,"title":163,"summary":164,"author":28},622,"2025-03-24","#OWL #AI Agent #MAS #MCP #CUA","article_res/cover/cb50ca7f2bf4d1ed50202d7406e1c19a.jpeg","article_res/cover/4aa7aa3badfacf3cc84121334f1050dd.jpeg","OWL: Multi-agent collaboration","OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation",{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},620,"2025-03-26","#LLM #Google #Gemini #AI Agent","article_res/cover/53751a6dbbe990b1eb0b63f3b062aed4.jpeg","article_res/cover/031344981f0a212ff82d1f3a64aa5756.jpeg","Gemini 2.5 Pro, claimed to be far ahead of the competition, has been released with great fanfare: comprehensively surpassing other LLMs and topping the global rankings","Gemini 2.5: Our most intelligent AI model",{"id":174,"publish_date":175,"is_original":23,"collection":176,"cover_url":177,"cover_url_1_1":178,"title":179,"summary":180,"author":28},616,"2025-03-29","#MAS #AI Agent #AI Coder #MetaGPT #MGX","article_res/cover/9dcd702ad2035902e5e77967c34a1f1e.jpeg","article_res/cover/0a97fc4a922753c8f46ff38792020df8.jpeg","MGX - An automated website-building platform composed of multiple AI Agents","Your 24/7 AI Team | Dream, Chat, Create.",{"title":182,"list":183},"OPENAI",[184,191,199,206],{"id":185,"publish_date":167,"is_original":23,"collection":186,"cover_url":187,"cover_url_1_1":188,"title":189,"summary":190,"author":28},619,"#OpenAI #AI Image Generator #4o #MMM #AR Transformer","article_res/cover/2faffc97fcecf3151552cb0fd3206d89.jpeg","article_res/cover/1133cb4948af44cee2e7fbe79efb69e5.jpeg","The native image function of GPT-4o is officially launched","Introducing 4o Image Generation",{"id":192,"publish_date":193,"is_original":4,"collection":194,"cover_url":195,"cover_url_1_1":196,"title":197,"summary":198,"author":28},434,"2023-07-15","#Anthropic #OpenAI #Google #AI Code Generator #Claude","article_res/cover/e1b6f600a2b9f262a4392684e5f2ce25.jpeg","article_res/cover/6e1772e83f78f9a351ab23d3e414adee.jpeg","Latest Updates on Google Bard /Anthropic Claude2 / ChatGPT Code Interpreter","We want our models to use their programming skills to provide more natural interfaces to the basic functions of our computers.  \n - OpenAI",{"id":200,"publish_date":201,"is_original":4,"collection":146,"cover_url":202,"cover_url_1_1":203,"title":204,"summary":205,"author":28},417,"2023-08-24","article_res/cover/bccf897d50a88b18364e35f7466387e0.jpeg","article_res/cover/2f871085c1073717c1703ae86e18056f.jpeg","The GPT-3.5 Turbo fine-tuning (fine-tuning function) has been released～","Developers can now bring their own data to customize GPT-3.5 Turbo for their use cases.",{"id":207,"publish_date":208,"is_original":4,"collection":209,"cover_url":210,"cover_url_1_1":211,"title":212,"summary":213,"author":28},407,"2023-09-22","#OpenAI #AI Image Generator","article_res/cover/c59005e903d35cfc32346e2756e2728a.jpeg","article_res/cover/ba011d265e6d84b5c8cb6fd6b757b6cc.jpeg","Dall-E 3","DALL·E 3 understands significantly more nuance and detail, allowing you to easily translate your ideas into images.",[215,221,241],{"title":10,"list":216},[217,218,219,220],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":223},"GOOGLE",[224,225,226,234],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},615,"2025-03-30","#AI Researcher #AI Science #HKU #Google #AI Agent","article_res/cover/21fadf906067714bb0db31ae13a77c15.jpeg","article_res/cover/2697999a72bd26b22e85f0e92936d3ed.jpeg","AI-Researcher: LLM-driven全自动 scientific research assistant","AI-Researcher: Fully-Automated Scientific Discovery with LLM Agents  \nOpen-Sourced Alternative to Google AI Co-Scientist",{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},463,"2023-05-09","article_res/cover/89800f207723acdb55fc53bf999ebdc9.jpeg","article_res/cover/5764f369b4accd8f83e94aa4c077a175.jpeg","The Smallville sandbox world - A town with 25 virtual residents","Believable proxies of human behavior can empower interactive apps: Immersive environment, Rehearsal space, Prototyping tool",{"title":242,"list":243},"NVIDIA",[],true,{"code":4,"msg":5,"data":246},{"id":88,"publish_date":89,"is_original":23,"collection":90,"articles_id":247,"cover_url":91,"cover_url_1_1":92,"title":93,"summary":94,"author":28,"content":248,"popular":249,"list":315,"category":371,"tag":372},"iV1AZN8KGywx2Mm1DVMJbQ","\u003Cdiv class=\"rich_media_content js_underline_content\n                       autoTypeSetting24psection\n            \" id=\"js_content\">\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;background-color: rgb(255, 255, 255);'>DisPose is a controllable human image animation method that improves video generation through motion field guidance and keypoint correspondence, developed in collaboration by several universities including Peking University, University of Science and Technology of China, Tsinghua University, and Hong Kong University of Science and Technology.\u003C/p>\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100009211\" data-ratio=\"0.38981481481481484\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" style=\"height: auto !important;\" src=\"./assets/17433488391620.03394232793180629.png\">\u003C/p>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">Introduction\u003C/span>\u003C/h3>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;background-color: rgb(255, 255, 255);'>Controllable human image animation aims to generate animated videos using reference images and driving videos. Recent studies have attempted to introduce additional dense conditions (such as depth maps) to ensure motion alignment due to the limited control signals provided by sparse guides like skeleton poses. However, when there are significant differences between the body shapes of the reference person and the driving video, such strict dense guidance can affect the quality of the generated video. This paper proposes DisPose, which aims to uncover more generalizable and effective control signals without requiring extra dense inputs. Specifically, it decouples sparse skeleton poses in human image animation into motion field guidance and keypoint correspondence.\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;background-color: rgb(255, 255, 255);'>Specifically, DisPose generates a dense motion field from sparse motion fields and reference images, providing regional-level guidance while maintaining the generalization ability of sparse pose control. Additionally, DisPose extracts diffusion features corresponding to pose keypoints from the reference image and transfers these point features to the target pose to provide unique identity information. To smoothly integrate into existing models, DisPose proposes a plug-and-play hybrid ControlNet that enhances the quality and consistency of generated videos while keeping the parameters of existing models frozen. Extensive qualitative and quantitative experiments demonstrate that DisPose has significant advantages over current methods.\u003C/p>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">\u003Cspan style='color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 20px;font-weight: 700;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);'>Examples\u003C/span>\u003C/span>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17433488377590.572806830599287.mp4\" poster=\"./assets/17433488361120.07409271374031379.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17433488371160.9048825405155472.mp4\" poster=\"./assets/17433488359070.26013739912652944.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Csection>\u003Cdiv style=\"height: 508px; background: rgb(0, 0, 0); border-radius: 4px; overflow: hidden; margin-bottom: 12px;\">\u003Cvideo src=\"./assets/17433488373110.5123531725798056.mp4\" poster=\"./assets/17433488359320.9194154101050527.jpeg\" controls=\"\" style=\"width: 100%;height: 100%;\">\u003C/video>\u003C/div>\u003C/section>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">\u003Cspan style='color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 20px;font-weight: 700;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);'>\u003C/span>\u003C/span>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">\u003Cbr>\u003C/span>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">Technical Framework\u003C/span>\u003C/h3>\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100009210\" data-ratio=\"0.5203703703703704\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" style=\"height: auto !important;\" src=\"./assets/17433488381610.6154444949511886.png\">\u003C/p>\u003Cp style='margin-bottom: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;padding-top: 8px;padding-bottom: 8px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;background-color: rgb(255, 255, 255);'>DisPose is a plug-and-play guidance module for decoupling pose guidance, extracting robust control signals from skeleton pose maps and reference images alone, without requiring additional dense inputs. Specifically, DisPose decouples pose guidance into motion field estimation and keypoint correspondence.\u003C/p>\u003Col style='margin-top: 8px;margin-bottom: 8px;cursor: pointer;padding-left: 25px;color: rgb(0, 0, 0);font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-size: 16px;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);' class=\"list-paddingleft-1\">\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">DisPose computes sparse motion fields using skeleton poses.\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">DisPose introduces reference-based dense motion fields, providing region-level motion signals via conditional motion propagation on the reference image. To enhance appearance consistency, DisPose extracts diffused features corresponding to keypoints in the reference image and transfers these point features to the target pose by calculating multi-scale point correspondences along motion trajectories. Architecturally, DisPose implements these decoupled control signals in a ControlNet-like manner, integrating them into existing methods.\u003C/section>\u003C/li>\u003Cli style=\"cursor: pointer;\">\u003Csection style=\"cursor: pointer;margin-top: 5px;margin-bottom: 5px;color: rgb(1, 1, 1);line-height: 1.8em;letter-spacing: 0em;\">The motion fields and point embeddings are injected into the latent video diffusion model, thereby generating accurate human image animations.\u003C/section>\u003C/li>\u003C/ol>\u003Ch3 style='margin-top: 30px;margin-bottom: 15px;color: rgba(0, 0, 0, 0.85);cursor: pointer;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;letter-spacing: normal;text-align: left;background-color: rgb(255, 255, 255);'>\u003Cspan style=\"cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;font-weight: bold;display: block;\">Contrast\u003C/span>\u003C/h3>\u003Cp style=\"text-align: center;\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100009212\" data-ratio=\"0.4166666666666667\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" style=\"height: auto !important;\" src=\"./assets/17433488388360.4649757708012967.png\">\u003C/p>\u003Cp style=\"display: none;\">\u003Cmp-style-type data-value=\"3\">\u003C/mp-style-type>\u003C/p>\u003C/div>",[250,259,267,275,282,289,298,306],{"id":251,"title_md5":252,"publish_date":253,"author_md5":254,"is_original":23,"collection":255,"summary_md5":256,"cover_url":257,"cover_url_1_1":258},119,"526d0f374ee04f855666789603e8bdf8","2024-11-25","bc27fa490c4d0d525bac812fc0793534","#AI Image Generator","f3521bc722ed915ad6e32225f2baf710","article_res/cover/621440bd2cd3230cd8481c4ed44378f1.jpeg","article_res/cover/177d7b243e807ca843f41ac82be538e1.jpeg",{"id":260,"title_md5":261,"publish_date":262,"author_md5":254,"is_original":23,"collection":263,"summary_md5":264,"cover_url":265,"cover_url_1_1":266},179,"c1319bc77ef0f21939e94d1a167944a4","2024-09-08","#AI Avatar #ByteDance","2fbad24b55db29159d47fdea5fe6ce3c","article_res/cover/4e2cf7ed5950551074b080d9d69c5c1f.jpeg","article_res/cover/53e3d3af86f031aa1eaf8e09dbf786ce.jpeg",{"id":268,"title_md5":269,"publish_date":270,"author_md5":271,"is_original":4,"collection":5,"summary_md5":272,"cover_url":273,"cover_url_1_1":274},598,"ccefbb37d14339b2167b7e38646e4faf","2022-03-19","8b3607d0f4181a3cb6ffdccf7185f09b","f4486bc5ab843101ef9327e41afc7967","article_res/cover/c0b3c15d1af9917e4888e30976e49606.jpeg","article_res/cover/9812ac9f3a9898718353dda47d933a29.jpeg",{"id":276,"title_md5":277,"publish_date":278,"author_md5":254,"is_original":4,"collection":5,"summary_md5":279,"cover_url":280,"cover_url_1_1":281},335,"6bd88f7d4a8878da48141da60737cdf5","2024-01-26","7e622ba5f679e76df22a989d0e87d851","article_res/cover/14bc34c5f592192af0f372971ce546cc.jpeg","article_res/cover/8b395057d151374561521132c9b01438.jpeg",{"id":283,"title_md5":284,"publish_date":285,"author_md5":254,"is_original":23,"collection":5,"summary_md5":286,"cover_url":287,"cover_url_1_1":288},403,"f7d3279a14cebc8c658a74eb5b0c82a7","2023-09-29","df72bde8ef86dd4939a679e18d44f613","article_res/cover/28fda6a29bd71e2f82e42574426d8290.jpeg","article_res/cover/90eb3f193393274e627cca44cbc3d64c.jpeg",{"id":290,"title_md5":291,"publish_date":292,"author_md5":293,"is_original":4,"collection":294,"summary_md5":295,"cover_url":296,"cover_url_1_1":297},588,"cddaa25577d3401f776d1bc2feff1bde","2022-03-29","70997a2931a60561e615f5171df5f9a5","#Philosophy","b13f09ac7bf387db64e2afa0882c7f63","article_res/cover/8a394d06bcf425e5d8f0c2b445313047.jpeg","article_res/cover/edded5e9be422882ea540097024968ff.jpeg",{"id":299,"title_md5":300,"publish_date":301,"author_md5":302,"is_original":23,"collection":5,"summary_md5":303,"cover_url":304,"cover_url_1_1":305},552,"4c6b59b807cd87944e25283c293a9bd9","2022-05-04","8936f53b23e9dfe2965d9e2ec23a4779","fd6e7c1af8ce6c247d8b12641f317b1d","article_res/cover/2b1d4e7150fc04a8a75f7214381d1a9c.jpeg","article_res/cover/1bcc7b2244f44a02ebad1fe9c0705d51.jpeg",{"id":307,"title_md5":308,"publish_date":309,"author_md5":310,"is_original":4,"collection":311,"summary_md5":312,"cover_url":313,"cover_url_1_1":314},165,"3182de43c64664bdac5e6cc02dce9af9","2024-09-29","fb0edf26dea7e6e8b89a99bc9d3a3170","#History of Intelligence #Neuroscience","8b51526e610bb04ea06ac18305a797ea","article_res/cover/8382e1be385e1572eea807d72618b064.jpeg","article_res/cover/8d840d1fb10cf0d9256b63524b17fb6f.jpeg",{"related":316,"small":356},[317,325,333,341,349],{"id":318,"publish_date":319,"is_original":23,"collection":320,"cover_url":321,"cover_url_1_1":322,"title":323,"summary":324,"author":28},16,"2025-03-06","#Google #AI Research #Gemini #AI Agents","article_res/cover/cd5eb106d2a8861222b43f0170e9e7b8.jpeg","article_res/cover/c28aa8bad4daf3b5927a38625d09d58b.jpeg","Google releases a new AI research assistant: AI co-scientist","Accelerating scientific breakthroughs with an AI co-scientist",{"id":326,"publish_date":327,"is_original":23,"collection":328,"cover_url":329,"cover_url_1_1":330,"title":331,"summary":332,"author":28},237,"2024-06-23","#AI Grant","article_res/cover/b1b8319c2d88f77217bf268133c0b58d.jpeg","article_res/cover/4c0f58fe75adf80ec93a69406fb5b234.jpeg","AI Grant Project List - Batch 1","Check out the thread to learn more about the companies",{"id":334,"publish_date":335,"is_original":23,"collection":336,"cover_url":337,"cover_url_1_1":338,"title":339,"summary":340,"author":28},48,"2025-02-05","#AI Agent #AI Avatar #Alibaba #AI Video Generator","article_res/cover/b4b09d99258fcd5045abbea17c3ad843.jpeg","article_res/cover/7d789b6343a2bdb2c8748571d9440ef1.jpeg","Alibaba's EMO2: Audio-Driven Talking Head Generation","EMO2: End-Effector Guided Audio-Driven Avatar Video Generation",{"id":342,"publish_date":343,"is_original":23,"collection":344,"cover_url":345,"cover_url_1_1":346,"title":347,"summary":348,"author":28},227,"2024-07-06","#Alibaba #AI Image Editor","article_res/cover/519d104c809b9a58349858c01411f1f3.jpeg","article_res/cover/4c1f3d58992d914e6d8c2bc0e56a7b1a.jpeg","MimicBrush: Zero-shot Image Editing and Reference Imitation","MimicBrush: Zero-shot Image Editing with Reference Imitation",{"id":350,"publish_date":351,"is_original":23,"collection":5,"cover_url":352,"cover_url_1_1":353,"title":354,"summary":355,"author":28},436,"2023-07-04","article_res/cover/74dac6e0da152222f9bcef4915543e43.jpeg","article_res/cover/24bd50763eb079f198644d4a39da6446.jpeg","Midjouney V5.2 Feature Introduction","This model produces more detailed, sharper results with better colors, contrast, and compositions.",[357,363,369],{"title":10,"list":358},[359,360,361,362],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":364},[365,366,367,368],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},{"title":242,"list":370},[],[8,9,10],[8,12,13,14,9,10,15,16,17,18],["Reactive",245],1754646407291]