[{"data":1,"prerenderedAt":373},["ShallowReactive",2],{"$fgukOamtKU1RtUiMFsqdObttmqPPQz0uc7bl_gj_LyX0":3,"$fcO1eMR7LJUKKeyOnFMrTInlHwr0kXbEkePUEz2xy6UQ":245,"article-425":372},{"code":4,"msg":5,"data":6},0,"",{"category":7,"tag":11,"popular":19,"latest":86,"banner":126,"list":151,"cache":244},[8,9,10],"Agent","OpenAI","LLM",[8,12,13,14,9,10,15,16,17,18],"Google","Nvidia","Claude","DeepSeek","OCR","Chat","Generator",[20,29,37,45,54,62,70,79],{"id":21,"publish_date":22,"is_original":23,"collection":5,"cover_url":24,"cover_url_1_1":25,"title":26,"summary":27,"author":28},411,"2023-09-10",1,"article_res/cover/451ef50c225a8dc61c4336506794d13b.jpeg","article_res/cover/3ba9dc7a72f87d40b20fc2d225289ee3.jpeg","Idealism","Reality is created by the mind, we can change our reality by changing our mind. - Plato","Renee's Entrepreneurial Journey",{"id":30,"publish_date":31,"is_original":23,"collection":32,"cover_url":33,"cover_url_1_1":34,"title":35,"summary":36,"author":28},108,"2024-12-07","#LLM #AGI #AI Agent","article_res/cover/0039044422e4ec9f61c18e8ee1693bb0.jpeg","article_res/cover/4220971b108a91d21407d87bb02fbaa6.jpeg","Freysa.ai: The World's First Adversarial AI Agent Game","说服 Freysa 把钱包里的钱都拿出来",{"id":38,"publish_date":39,"is_original":23,"collection":40,"cover_url":41,"cover_url_1_1":42,"title":43,"summary":44,"author":28},12,"2025-03-09","#Oxford #Reasoning #LLM #Tool Use","article_res/cover/d448e9b3617a0b5302e1bd10c438bca9.jpeg","article_res/cover/864a468f9cc4c9317efadb3811909888.jpeg","Agentic Reasoning Framework - Significantly enhance the reasoning ability of LLMs through the integration of external tools using agents","Agentic Reasoning: Reasoning LLMs with Tools for Deep Research",{"id":46,"publish_date":47,"is_original":4,"collection":48,"cover_url":49,"cover_url_1_1":50,"title":51,"summary":52,"author":53},480,"2023-04-14","#Stable Diffusion","article_res/cover/0bdbe7cb1de4a78e54536e5d9afa7ec9.jpeg","article_res/cover/b3d6ffec0608dcfaf18c5a69906d1490.jpeg","【AIGC Learning】Generate Prompts Using Word Graphs - Stable Diffusion Web UI Series 13","AI will become a powerful tool in education, transforming the way we learn and deliver instruction.  \n- Reid Hoffman","--",{"id":55,"publish_date":56,"is_original":4,"collection":57,"cover_url":58,"cover_url_1_1":59,"title":60,"summary":61,"author":28},413,"2023-09-08","#Neuroscience","article_res/cover/74f8302d78a23d9430f22171eae136b6.jpeg","article_res/cover/87ca08af81bb304746be5261160964c0.jpeg","Can machines be conscious?","Do we have an ethical obligation to not turn off conscious machines? Would turning them off be murder? No. I don't lose any sleep over unplugging a conscious machine.\n- Jeff Hawkins, \"A Thousand Brains\"",{"id":63,"publish_date":64,"is_original":23,"collection":65,"cover_url":66,"cover_url_1_1":67,"title":68,"summary":69,"author":28},178,"2024-09-09","#Entrepreneurship","article_res/cover/a7224f025b55d1820408085faef63079.jpeg","article_res/cover/11a9995b096cbf64465ef01b8673b154.jpeg","37signals company","This damn sense of relaxation",{"id":71,"publish_date":72,"is_original":4,"collection":73,"cover_url":74,"cover_url_1_1":75,"title":76,"summary":77,"author":78},460,"2023-05-12","#Google","article_res/cover/b970687b12faa52da976f91248c2aa7b.jpeg","article_res/cover/d1e71b52cfd2c63bc6e71f3e85ff135c.jpeg","Learn what BRC-20 and Ordinals are using Google Bard","Ordinals - a new protocol that allows users to store arbitrary data on the Bitcoin blockchain","Google Bard mainly writes",{"id":80,"publish_date":81,"is_original":23,"collection":5,"cover_url":82,"cover_url_1_1":83,"title":84,"summary":85,"author":28},309,"2024-03-26","article_res/cover/9877f95894ee88532d0e6012c23a2df3.jpeg","article_res/cover/20092164ddc109ce6ae56b1984246751.jpeg","Learning the Cancun Upgrade with lepton and perplexity","Building a quick conversation-based search demo with Lepton AI.",[87,95,103,111,119],{"id":88,"publish_date":89,"is_original":23,"collection":90,"cover_url":91,"cover_url_1_1":92,"title":93,"summary":94,"author":28},627,"2025-03-20","#AI Avatar #AI Video Generation","article_res/cover/d95481358f73924989f8c4ee9c75d1c8.jpeg","article_res/cover/b74bc0fab01f8b6a6aa87696c0c3ed8b.jpeg","DisPose: Generating Animated Videos by Driving Video with Reference Images","DisPose is a controllable human image animation method that enhances video generation.",{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},626,"2025-03-21","#Deep Dive into LLMs #LLM #RL #Andrej Karpathy #AlphaGo","article_res/cover/446553a5c8f8f2f07d97b20eaee84e56.jpeg","article_res/cover/e6c2823409c9b34624064b9acbaca6f1.jpeg","AlphaGo and the Power of Reinforcement Learning - Andrej Karpathy's Deep Dive on LLMs (Part 9)","Simply learning from humans will never surpass human capabilities.",{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},625,"2025-03-22","#Deep Dive into LLMs #LLM #RL #RLHF #Andrej Karpathy","article_res/cover/8da81d38b1e5cf558a164710fd8a5389.jpeg","article_res/cover/96f028d76c362a99a0dd56389e8f7a9b.jpeg","Reinforcement Learning from Human Feedback (RLHF) - Andrej Karpathy's Deep Dive on LLMs (Part 10)","Fine-Tuning Language Models from Human Preferences",{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},624,"2025-03-23","#Deep Dive into LLMs #LLM #Andrej Karpathy #AI Agent #MMM","article_res/cover/a5e7c3d48bb09109684d6513287c661d.jpeg","article_res/cover/d3f22b7c0ab8d82fd2da457a299e0773.jpeg","The Future of Large Language Models - Andrej Karpathy's In-Depth Explanation of LLM (Part 11)","preview of things to come",{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},623,"#Google #Voe #AI Video Generation","article_res/cover/c44062fea0f336c2b96b3928292392c2.jpeg","article_res/cover/a041041c69092ad3db191c5bf3ff981b.jpeg","Trial of Google's video generation model VOE2","Our state-of-the-art video generation model",[127,135,143],{"id":128,"publish_date":129,"is_original":23,"collection":130,"cover_url":131,"cover_url_1_1":132,"title":133,"summary":134,"author":28},300,"2024-04-16","#AI in Science #AGI","article_res/cover/6bf01e793e0f33e848572412eebdf9b0.jpeg","article_res/cover/91a5ee21dafecb914fabeb9430d46ec1.jpeg","Would Einstein lose his job - AI and Quantum Computing: A Glimpse into the Near Future","So Einstein's job is still safe.",{"id":136,"publish_date":137,"is_original":23,"collection":138,"cover_url":139,"cover_url_1_1":140,"title":141,"summary":142,"author":28},101,"2024-12-14","#Nvidia #AI 3D Generator","article_res/cover/693e07c85980c5c0c8fde3f037733f23.jpeg","article_res/cover/9ea8edff2d5d303ff3fffff3f6f9c3d9.jpeg","NVIDIA's open-source 3D project LLaMA-Mesh","LLaMA-Mesh: Unifying 3D Mesh Generation with Language Models",{"id":144,"publish_date":145,"is_original":23,"collection":146,"cover_url":147,"cover_url_1_1":148,"title":149,"summary":150,"author":28},131,"2024-11-10","#OpenAI","article_res/cover/87f8ed353ce39f31960e7cdfaf075a35.jpeg","article_res/cover/f597a63935f5cd32e484b4aadd6019e8.jpeg","ChatGPT has launched the Search function","Get fast, timely answers with links to relevant web sources.",{"big":152,"small":214},[153,181],{"title":154,"list":155},"AGENT",[156,157,165,173],{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":158,"publish_date":159,"is_original":23,"collection":160,"cover_url":161,"cover_url_1_1":162,"title":163,"summary":164,"author":28},622,"2025-03-24","#OWL #AI Agent #MAS #MCP #CUA","article_res/cover/cb50ca7f2bf4d1ed50202d7406e1c19a.jpeg","article_res/cover/4aa7aa3badfacf3cc84121334f1050dd.jpeg","OWL: Multi-agent collaboration","OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation",{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},620,"2025-03-26","#LLM #Google #Gemini #AI Agent","article_res/cover/53751a6dbbe990b1eb0b63f3b062aed4.jpeg","article_res/cover/031344981f0a212ff82d1f3a64aa5756.jpeg","Gemini 2.5 Pro, claimed to be far ahead of the competition, has been released with great fanfare: comprehensively surpassing other LLMs and topping the global rankings","Gemini 2.5: Our most intelligent AI model",{"id":174,"publish_date":175,"is_original":23,"collection":176,"cover_url":177,"cover_url_1_1":178,"title":179,"summary":180,"author":28},616,"2025-03-29","#MAS #AI Agent #AI Coder #MetaGPT #MGX","article_res/cover/9dcd702ad2035902e5e77967c34a1f1e.jpeg","article_res/cover/0a97fc4a922753c8f46ff38792020df8.jpeg","MGX - An automated website-building platform composed of multiple AI Agents","Your 24/7 AI Team | Dream, Chat, Create.",{"title":182,"list":183},"OPENAI",[184,191,199,206],{"id":185,"publish_date":167,"is_original":23,"collection":186,"cover_url":187,"cover_url_1_1":188,"title":189,"summary":190,"author":28},619,"#OpenAI #AI Image Generator #4o #MMM #AR Transformer","article_res/cover/2faffc97fcecf3151552cb0fd3206d89.jpeg","article_res/cover/1133cb4948af44cee2e7fbe79efb69e5.jpeg","The native image function of GPT-4o is officially launched","Introducing 4o Image Generation",{"id":192,"publish_date":193,"is_original":4,"collection":194,"cover_url":195,"cover_url_1_1":196,"title":197,"summary":198,"author":28},434,"2023-07-15","#Anthropic #OpenAI #Google #AI Code Generator #Claude","article_res/cover/e1b6f600a2b9f262a4392684e5f2ce25.jpeg","article_res/cover/6e1772e83f78f9a351ab23d3e414adee.jpeg","Latest Updates on Google Bard /Anthropic Claude2 / ChatGPT Code Interpreter","We want our models to use their programming skills to provide more natural interfaces to the basic functions of our computers.  \n - OpenAI",{"id":200,"publish_date":201,"is_original":4,"collection":146,"cover_url":202,"cover_url_1_1":203,"title":204,"summary":205,"author":28},417,"2023-08-24","article_res/cover/bccf897d50a88b18364e35f7466387e0.jpeg","article_res/cover/2f871085c1073717c1703ae86e18056f.jpeg","The GPT-3.5 Turbo fine-tuning (fine-tuning function) has been released～","Developers can now bring their own data to customize GPT-3.5 Turbo for their use cases.",{"id":207,"publish_date":208,"is_original":4,"collection":209,"cover_url":210,"cover_url_1_1":211,"title":212,"summary":213,"author":28},407,"2023-09-22","#OpenAI #AI Image Generator","article_res/cover/c59005e903d35cfc32346e2756e2728a.jpeg","article_res/cover/ba011d265e6d84b5c8cb6fd6b757b6cc.jpeg","Dall-E 3","DALL·E 3 understands significantly more nuance and detail, allowing you to easily translate your ideas into images.",[215,221,241],{"title":10,"list":216},[217,218,219,220],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":223},"GOOGLE",[224,225,226,234],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},615,"2025-03-30","#AI Researcher #AI Science #HKU #Google #AI Agent","article_res/cover/21fadf906067714bb0db31ae13a77c15.jpeg","article_res/cover/2697999a72bd26b22e85f0e92936d3ed.jpeg","AI-Researcher: LLM-driven全自动 scientific research assistant","AI-Researcher: Fully-Automated Scientific Discovery with LLM Agents  \nOpen-Sourced Alternative to Google AI Co-Scientist",{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},463,"2023-05-09","article_res/cover/89800f207723acdb55fc53bf999ebdc9.jpeg","article_res/cover/5764f369b4accd8f83e94aa4c077a175.jpeg","The Smallville sandbox world - A town with 25 virtual residents","Believable proxies of human behavior can empower interactive apps: Immersive environment, Rehearsal space, Prototyping tool",{"title":242,"list":243},"NVIDIA",[],true,{"code":4,"msg":5,"data":246},{"id":247,"publish_date":248,"is_original":4,"collection":249,"articles_id":250,"cover_url":251,"cover_url_1_1":252,"title":253,"summary":254,"author":28,"content":255,"popular":256,"list":320,"category":370,"tag":371},425,"2023-08-04","#AI Agents #AI Game #AI Code Generator #AI Agent","jSPWdzbQRF0MBSHU1ujVKg","article_res/cover/e364b56ed639fae29cdde957447165e4.jpeg","article_res/cover/3441f67d984d02d4373fbb45770a22f1.jpeg","ChatDev - AI Agents for Game Development","Collaboration allows us to know more than we are capable of knowing by ourselves.  - Paul Solarz","\u003Cdiv class=\"rich_media_content js_underline_content\n                       defaultNoSetting\n            \" id=\"js_content\">\u003Csection>\u003Csection style=\"display: inline-block;\">\u003Cimg data-ratio=\"0.9558587479935795\" data-type=\"jpg\" data-w=\"1246\" src=\"./assets/17434959773920.4240014344587626.jpeg\">\u003C/section>​\u003C/section>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Cp style=\"box-sizing: border-box;margin: 0px 0px 1.2em;\">It has always been said that Agent is the way to AGI. Many people talked about Tsinghua's ChatDev before, and today I will learn about it.\u003C/p>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">Software engineering is a field characterized by complex decision-making processes, often relying on subtle intuition and consultation. Recent advances in deep learning have begun to revolutionize software engineering practices by implementing refined designs at various stages of software development. In this paper, we propose an innovative paradigm that leverages large language models (LLMs) throughout the entire software development process, simplifying and unifying key processes through natural language communication, thus eliminating the need for specialized models at each stage. At the core of this paradigm is CHATDEV, a virtual chat-powered software development company that meticulously divides the development process into four distinct phases: design, coding, testing, and documentation writing. Each phase involves the participation of a team, such as programmers, code reviewers, and test engineers, promoting collaborative dialogue and facilitating smooth workflows. The chat chain acts as a facilitator, breaking down each phase into atomic subtasks. This enables dual roles, proposing and verifying solutions through context-sensitive communication, thereby efficiently solving specific subtasks. A significant analysis of CHATDEV highlights its remarkable efficacy in software generation, enabling the entire software development process to be completed in less than seven minutes at a cost of less than one dollar. It not only identifies and mitigates potential vulnerabilities but also corrects potential misconceptions while maintaining commendable efficiency and cost-effectiveness. The potential of CHATDEV reveals new possibilities for integrating LLMs into the field of software development.\u003C/p>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Cblockquote style=\"box-sizing: border-box;margin: 1.2em 0px;color: rgba(0, 0, 0, 0.5);padding-left: 1.5em;border-left-width: 5px;border-left-style: solid;border-left-color: rgba(0, 0, 0, 0.1);\">\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">Collaboration allows us to know more than we are capable of knowing by ourselves. It empowers us to think differently, access information we wouldn’t have otherwise, and combine ideas as we work together towards a shared goal.\u003C/p>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">— Paul Solarz\u003C/p>\u003C/blockquote>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Csection>\u003Csection style=\"display: inline-block;\">\u003Cimg data-ratio=\"0.35144927536231885\" data-type=\"jpg\" data-w=\"1932\" src=\"./assets/17434959779210.2091510969050654.jpeg\">\u003C/section>​\u003C/section>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">Different agents play different experts and then collaborate to complete the development of a game. Communication occurs between different roles, with added reflection phases.\u003C/p>\u003Csection>\u003Csection style=\"display: inline-block;\">\u003Cimg data-ratio=\"0.3710265763418447\" data-type=\"jpg\" data-w=\"1919\" src=\"./assets/17434959779210.34847028794810986.jpeg\">\u003C/section>​\u003C/section>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">\u003Cbr>\u003C/p>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Cul style=\"box-sizing: border-box;margin: 1.2em 0px;\">\u003Cli style=\"box-sizing: border-box;\">The \"memory flow\" will save records of each robot's conversation in each round so they can be read at any time, ensuring continuity of thought.\u003C/li>\u003Cli style=\"box-sizing: border-box;\">The \"self-reflection\" mechanism generates a \"pseudo-self\" and feeds back the problems and related dialogues to the instructor when the robot completes its respective tasks but fails to meet the requirements.\u003C/li>\u003C/ul>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Csection>\u003Csection style=\"display: inline-block;\">\u003Cimg data-ratio=\"0.2464751958224543\" data-type=\"jpg\" data-w=\"1915\" src=\"./assets/17434959778800.7873759572373435.jpeg\">\u003C/section>​\u003C/section>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">However, only the paper has been released, without open-sourcing the code or making the demo available, so I can't reproduce it.\u003C/p>\u003C/div>\u003Cdiv class=\"cl-preview-section\" style='box-sizing: border-box;caret-color: rgba(0, 0, 0, 0.75);color: rgba(0, 0, 0, 0.75);font-family: Lato, \"Helvetica Neue\", Helvetica, sans-serif;font-size: 17px;font-style: normal;font-variant-caps: normal;font-weight: 400;letter-spacing: normal;orphans: auto;text-align: start;text-indent: 0px;text-transform: none;white-space: normal;widows: auto;word-spacing: 0px;-webkit-tap-highlight-color: rgba(26, 26, 26, 0.3);-webkit-text-stroke-width: 0px;text-decoration: none;'>\u003Cp style=\"box-sizing: border-box;margin: 1.2em 0px;\">Paper link: https://arxiv.org/abs/2307.07924\u003C/p>\u003C/div>\u003Cp style=\"display: none;\">\u003Cmp-style-type data-value=\"10000\">\u003C/mp-style-type>\u003C/p>\u003C/div>",[257,266,273,281,289,296,304,312],{"id":258,"title_md5":259,"publish_date":260,"author_md5":261,"is_original":23,"collection":262,"summary_md5":263,"cover_url":264,"cover_url_1_1":265},27,"2d570a836477d17b7cb92d3363e42583","2025-02-25","bc27fa490c4d0d525bac812fc0793534","#Claude #Anthropic #LLM #Think","fb9394f0ea7dff89616cfd94fc8a31c1","article_res/cover/cb055a3950605a1a7ba7163daa89f267.jpeg","article_res/cover/7533a64e36e80a49011564195a9d03fa.jpeg",{"id":267,"title_md5":268,"publish_date":269,"author_md5":261,"is_original":4,"collection":5,"summary_md5":270,"cover_url":271,"cover_url_1_1":272},348,"91e94b13e4faceb3659f55c0b8888205","2024-01-08","aba3044ccaa261665de2c8f201f67eec","article_res/cover/76e6110abbe80f7fa055e94ebb418eb2.jpeg","article_res/cover/2ca21b605ea966964f657a58c4e05f93.jpeg",{"id":274,"title_md5":275,"publish_date":276,"author_md5":277,"is_original":23,"collection":65,"summary_md5":278,"cover_url":279,"cover_url_1_1":280},548,"fd2f666a2c6e1f073d210cbeedf4e855","2022-05-08","8936f53b23e9dfe2965d9e2ec23a4779","a41a2645c872279ec2902aab727bfeab","article_res/cover/57ff459a7216dda5ec0ebf150d65a192.jpeg","article_res/cover/027261d0f3d4ff543ac27ee3dacbee66.jpeg",{"id":282,"title_md5":283,"publish_date":284,"author_md5":261,"is_original":23,"collection":285,"summary_md5":286,"cover_url":287,"cover_url_1_1":288},212,"4ce5c798f705e7c1f5099b03bea44a63","2024-07-23","#AI Virtual Try-On #Tencent","ce21ffcdd28f0fd9821aab5ec19a545c","article_res/cover/4b843969ae2bdf19e55f75c824be7d4f.jpeg","article_res/cover/762ea863dd1d1fa32990288b476313c5.jpeg",{"id":290,"title_md5":291,"publish_date":292,"author_md5":261,"is_original":23,"collection":146,"summary_md5":293,"cover_url":294,"cover_url_1_1":295},280,"455091bdbdbee638b3a321df9edce155","2024-05-09","a795ddf3415ce755e488296baf27c81d","article_res/cover/be0bb529e4f8f58b671ad675b89e31ec.jpeg","article_res/cover/d9f9a24f6bf2aa07830b5997f128de0c.jpeg",{"id":297,"title_md5":298,"publish_date":299,"author_md5":261,"is_original":23,"collection":300,"summary_md5":301,"cover_url":302,"cover_url_1_1":303},112,"697f5512b716dc93c3f6d446e2968edd","2024-12-03","#AI Code Generator","a3245816e1c6b3fa972cb99a56871344","article_res/cover/42bf26f9f2ca5af22ef473673cb9f518.jpeg","article_res/cover/9aaa9c965e49781053a5e7aeb3491173.jpeg",{"id":305,"title_md5":306,"publish_date":307,"author_md5":308,"is_original":4,"collection":5,"summary_md5":309,"cover_url":310,"cover_url_1_1":311},495,"9b15ccc6efafbde434556a70af4440d0","2023-03-30","9f3428c4d2d88afac6a16510d115e41c","8c3a91662be7ddf873148e266df78686","article_res/cover/9cc1e0b26ea462d3277b0d98b9a74620.jpeg","article_res/cover/8ecf7bdc377a506ebe83640c883cdff7.jpeg",{"id":313,"title_md5":314,"publish_date":315,"author_md5":261,"is_original":23,"collection":316,"summary_md5":317,"cover_url":318,"cover_url_1_1":319},34,"5084f00a10acd3dbc6a50ddd42d77466","2025-02-19","#LLM #Grok3 #DeepSeek #ChatGPT #Think","198e8bc111a30d576f1929611265473f","article_res/cover/825ed24949e3d283bf6ddb0e224023bd.jpeg","article_res/cover/2a174a1c0753acd1cc25a4808c5e589c.jpeg",{"related":321,"small":355},[322,330,338,339,347],{"id":323,"publish_date":324,"is_original":23,"collection":325,"cover_url":326,"cover_url_1_1":327,"title":328,"summary":329,"author":28},135,"2024-11-06","#Philosophy #Psychology","article_res/cover/89c64742d51a90a947f4329bb25d9ea6.jpeg","article_res/cover/7c0d4dae638efc6c73a6ae53a8ed1c58.jpeg","How to continuously expand the boundaries of cognition","Too often, a vast collection of possessions ends up possessing its owner.",{"id":331,"publish_date":332,"is_original":23,"collection":333,"cover_url":334,"cover_url_1_1":335,"title":336,"summary":337,"author":28},80,"2025-01-04","#Google #AI Video Editor","article_res/cover/244ba813af9084ab8b2fb42748da185d.jpeg","article_res/cover/4af914a0a7176eac1a45541f144a623a.jpeg","Google's Motion Prompting Controls Video Generation Through Motion Trajectories","Motion Prompting: Controlling Video Generation with Motion Trajectories",{"id":21,"publish_date":22,"is_original":23,"collection":5,"cover_url":24,"cover_url_1_1":25,"title":26,"summary":27,"author":28},{"id":340,"publish_date":341,"is_original":23,"collection":342,"cover_url":343,"cover_url_1_1":344,"title":345,"summary":346,"author":28},146,"2024-10-23","#Anthropic #Claude","article_res/cover/31731a776b0f2721f65180b964fec1c2.jpeg","article_res/cover/780acc08a7d276a08659c9f2892cb854.jpeg","Anthropic's Claude can directly operate a computer","Developing a computer use model",{"id":348,"publish_date":349,"is_original":23,"collection":350,"cover_url":351,"cover_url_1_1":352,"title":353,"summary":354,"author":28},352,"2024-01-03","#AGI #AI Agent","article_res/cover/5a90dc2f33c6472e243ddf7ba925e2d1.jpeg","article_res/cover/6064536658bfbf1176aa0fa3c5fe640e.jpeg","AI Agent Application Market Map","Autonomous agents have long been a prominent research focus in both academic and industry communities",[356,362,368],{"title":10,"list":357},[358,359,360,361],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":363},[364,365,366,367],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},{"title":242,"list":369},[],[8,9,10],[8,12,13,14,9,10,15,16,17,18],["Reactive",245],1754646411306]