[{"data":1,"prerenderedAt":373},["ShallowReactive",2],{"$fgukOamtKU1RtUiMFsqdObttmqPPQz0uc7bl_gj_LyX0":3,"$fPEiyKQWulb-besG5MHj6EpUksApaLkGt23rU5wcB_7Q":245,"article-4":372},{"code":4,"msg":5,"data":6},0,"",{"category":7,"tag":11,"popular":19,"latest":86,"banner":126,"list":151,"cache":244},[8,9,10],"Agent","OpenAI","LLM",[8,12,13,14,9,10,15,16,17,18],"Google","Nvidia","Claude","DeepSeek","OCR","Chat","Generator",[20,29,37,45,54,62,70,79],{"id":21,"publish_date":22,"is_original":23,"collection":5,"cover_url":24,"cover_url_1_1":25,"title":26,"summary":27,"author":28},411,"2023-09-10",1,"article_res/cover/451ef50c225a8dc61c4336506794d13b.jpeg","article_res/cover/3ba9dc7a72f87d40b20fc2d225289ee3.jpeg","Idealism","Reality is created by the mind, we can change our reality by changing our mind. - Plato","Renee's Entrepreneurial Journey",{"id":30,"publish_date":31,"is_original":23,"collection":32,"cover_url":33,"cover_url_1_1":34,"title":35,"summary":36,"author":28},108,"2024-12-07","#LLM #AGI #AI Agent","article_res/cover/0039044422e4ec9f61c18e8ee1693bb0.jpeg","article_res/cover/4220971b108a91d21407d87bb02fbaa6.jpeg","Freysa.ai: The World's First Adversarial AI Agent Game","说服 Freysa 把钱包里的钱都拿出来",{"id":38,"publish_date":39,"is_original":23,"collection":40,"cover_url":41,"cover_url_1_1":42,"title":43,"summary":44,"author":28},12,"2025-03-09","#Oxford #Reasoning #LLM #Tool Use","article_res/cover/d448e9b3617a0b5302e1bd10c438bca9.jpeg","article_res/cover/864a468f9cc4c9317efadb3811909888.jpeg","Agentic Reasoning Framework - Significantly enhance the reasoning ability of LLMs through the integration of external tools using agents","Agentic Reasoning: Reasoning LLMs with Tools for Deep Research",{"id":46,"publish_date":47,"is_original":4,"collection":48,"cover_url":49,"cover_url_1_1":50,"title":51,"summary":52,"author":53},480,"2023-04-14","#Stable Diffusion","article_res/cover/0bdbe7cb1de4a78e54536e5d9afa7ec9.jpeg","article_res/cover/b3d6ffec0608dcfaf18c5a69906d1490.jpeg","【AIGC Learning】Generate Prompts Using Word Graphs - Stable Diffusion Web UI Series 13","AI will become a powerful tool in education, transforming the way we learn and deliver instruction.  \n- Reid Hoffman","--",{"id":55,"publish_date":56,"is_original":4,"collection":57,"cover_url":58,"cover_url_1_1":59,"title":60,"summary":61,"author":28},413,"2023-09-08","#Neuroscience","article_res/cover/74f8302d78a23d9430f22171eae136b6.jpeg","article_res/cover/87ca08af81bb304746be5261160964c0.jpeg","Can machines be conscious?","Do we have an ethical obligation to not turn off conscious machines? Would turning them off be murder? No. I don't lose any sleep over unplugging a conscious machine.\n- Jeff Hawkins, \"A Thousand Brains\"",{"id":63,"publish_date":64,"is_original":23,"collection":65,"cover_url":66,"cover_url_1_1":67,"title":68,"summary":69,"author":28},178,"2024-09-09","#Entrepreneurship","article_res/cover/a7224f025b55d1820408085faef63079.jpeg","article_res/cover/11a9995b096cbf64465ef01b8673b154.jpeg","37signals company","This damn sense of relaxation",{"id":71,"publish_date":72,"is_original":4,"collection":73,"cover_url":74,"cover_url_1_1":75,"title":76,"summary":77,"author":78},460,"2023-05-12","#Google","article_res/cover/b970687b12faa52da976f91248c2aa7b.jpeg","article_res/cover/d1e71b52cfd2c63bc6e71f3e85ff135c.jpeg","Learn what BRC-20 and Ordinals are using Google Bard","Ordinals - a new protocol that allows users to store arbitrary data on the Bitcoin blockchain","Google Bard mainly writes",{"id":80,"publish_date":81,"is_original":23,"collection":5,"cover_url":82,"cover_url_1_1":83,"title":84,"summary":85,"author":28},309,"2024-03-26","article_res/cover/9877f95894ee88532d0e6012c23a2df3.jpeg","article_res/cover/20092164ddc109ce6ae56b1984246751.jpeg","Learning the Cancun Upgrade with lepton and perplexity","Building a quick conversation-based search demo with Lepton AI.",[87,95,103,111,119],{"id":88,"publish_date":89,"is_original":23,"collection":90,"cover_url":91,"cover_url_1_1":92,"title":93,"summary":94,"author":28},627,"2025-03-20","#AI Avatar #AI Video Generation","article_res/cover/d95481358f73924989f8c4ee9c75d1c8.jpeg","article_res/cover/b74bc0fab01f8b6a6aa87696c0c3ed8b.jpeg","DisPose: Generating Animated Videos by Driving Video with Reference Images","DisPose is a controllable human image animation method that enhances video generation.",{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},626,"2025-03-21","#Deep Dive into LLMs #LLM #RL #Andrej Karpathy #AlphaGo","article_res/cover/446553a5c8f8f2f07d97b20eaee84e56.jpeg","article_res/cover/e6c2823409c9b34624064b9acbaca6f1.jpeg","AlphaGo and the Power of Reinforcement Learning - Andrej Karpathy's Deep Dive on LLMs (Part 9)","Simply learning from humans will never surpass human capabilities.",{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},625,"2025-03-22","#Deep Dive into LLMs #LLM #RL #RLHF #Andrej Karpathy","article_res/cover/8da81d38b1e5cf558a164710fd8a5389.jpeg","article_res/cover/96f028d76c362a99a0dd56389e8f7a9b.jpeg","Reinforcement Learning from Human Feedback (RLHF) - Andrej Karpathy's Deep Dive on LLMs (Part 10)","Fine-Tuning Language Models from Human Preferences",{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},624,"2025-03-23","#Deep Dive into LLMs #LLM #Andrej Karpathy #AI Agent #MMM","article_res/cover/a5e7c3d48bb09109684d6513287c661d.jpeg","article_res/cover/d3f22b7c0ab8d82fd2da457a299e0773.jpeg","The Future of Large Language Models - Andrej Karpathy's In-Depth Explanation of LLM (Part 11)","preview of things to come",{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},623,"#Google #Voe #AI Video Generation","article_res/cover/c44062fea0f336c2b96b3928292392c2.jpeg","article_res/cover/a041041c69092ad3db191c5bf3ff981b.jpeg","Trial of Google's video generation model VOE2","Our state-of-the-art video generation model",[127,135,143],{"id":128,"publish_date":129,"is_original":23,"collection":130,"cover_url":131,"cover_url_1_1":132,"title":133,"summary":134,"author":28},300,"2024-04-16","#AI in Science #AGI","article_res/cover/6bf01e793e0f33e848572412eebdf9b0.jpeg","article_res/cover/91a5ee21dafecb914fabeb9430d46ec1.jpeg","Would Einstein lose his job - AI and Quantum Computing: A Glimpse into the Near Future","So Einstein's job is still safe.",{"id":136,"publish_date":137,"is_original":23,"collection":138,"cover_url":139,"cover_url_1_1":140,"title":141,"summary":142,"author":28},101,"2024-12-14","#Nvidia #AI 3D Generator","article_res/cover/693e07c85980c5c0c8fde3f037733f23.jpeg","article_res/cover/9ea8edff2d5d303ff3fffff3f6f9c3d9.jpeg","NVIDIA's open-source 3D project LLaMA-Mesh","LLaMA-Mesh: Unifying 3D Mesh Generation with Language Models",{"id":144,"publish_date":145,"is_original":23,"collection":146,"cover_url":147,"cover_url_1_1":148,"title":149,"summary":150,"author":28},131,"2024-11-10","#OpenAI","article_res/cover/87f8ed353ce39f31960e7cdfaf075a35.jpeg","article_res/cover/f597a63935f5cd32e484b4aadd6019e8.jpeg","ChatGPT has launched the Search function","Get fast, timely answers with links to relevant web sources.",{"big":152,"small":214},[153,181],{"title":154,"list":155},"AGENT",[156,157,165,173],{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":158,"publish_date":159,"is_original":23,"collection":160,"cover_url":161,"cover_url_1_1":162,"title":163,"summary":164,"author":28},622,"2025-03-24","#OWL #AI Agent #MAS #MCP #CUA","article_res/cover/cb50ca7f2bf4d1ed50202d7406e1c19a.jpeg","article_res/cover/4aa7aa3badfacf3cc84121334f1050dd.jpeg","OWL: Multi-agent collaboration","OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation",{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},620,"2025-03-26","#LLM #Google #Gemini #AI Agent","article_res/cover/53751a6dbbe990b1eb0b63f3b062aed4.jpeg","article_res/cover/031344981f0a212ff82d1f3a64aa5756.jpeg","Gemini 2.5 Pro, claimed to be far ahead of the competition, has been released with great fanfare: comprehensively surpassing other LLMs and topping the global rankings","Gemini 2.5: Our most intelligent AI model",{"id":174,"publish_date":175,"is_original":23,"collection":176,"cover_url":177,"cover_url_1_1":178,"title":179,"summary":180,"author":28},616,"2025-03-29","#MAS #AI Agent #AI Coder #MetaGPT #MGX","article_res/cover/9dcd702ad2035902e5e77967c34a1f1e.jpeg","article_res/cover/0a97fc4a922753c8f46ff38792020df8.jpeg","MGX - An automated website-building platform composed of multiple AI Agents","Your 24/7 AI Team | Dream, Chat, Create.",{"title":182,"list":183},"OPENAI",[184,191,199,206],{"id":185,"publish_date":167,"is_original":23,"collection":186,"cover_url":187,"cover_url_1_1":188,"title":189,"summary":190,"author":28},619,"#OpenAI #AI Image Generator #4o #MMM #AR Transformer","article_res/cover/2faffc97fcecf3151552cb0fd3206d89.jpeg","article_res/cover/1133cb4948af44cee2e7fbe79efb69e5.jpeg","The native image function of GPT-4o is officially launched","Introducing 4o Image Generation",{"id":192,"publish_date":193,"is_original":4,"collection":194,"cover_url":195,"cover_url_1_1":196,"title":197,"summary":198,"author":28},434,"2023-07-15","#Anthropic #OpenAI #Google #AI Code Generator #Claude","article_res/cover/e1b6f600a2b9f262a4392684e5f2ce25.jpeg","article_res/cover/6e1772e83f78f9a351ab23d3e414adee.jpeg","Latest Updates on Google Bard /Anthropic Claude2 / ChatGPT Code Interpreter","We want our models to use their programming skills to provide more natural interfaces to the basic functions of our computers.  \n - OpenAI",{"id":200,"publish_date":201,"is_original":4,"collection":146,"cover_url":202,"cover_url_1_1":203,"title":204,"summary":205,"author":28},417,"2023-08-24","article_res/cover/bccf897d50a88b18364e35f7466387e0.jpeg","article_res/cover/2f871085c1073717c1703ae86e18056f.jpeg","The GPT-3.5 Turbo fine-tuning (fine-tuning function) has been released～","Developers can now bring their own data to customize GPT-3.5 Turbo for their use cases.",{"id":207,"publish_date":208,"is_original":4,"collection":209,"cover_url":210,"cover_url_1_1":211,"title":212,"summary":213,"author":28},407,"2023-09-22","#OpenAI #AI Image Generator","article_res/cover/c59005e903d35cfc32346e2756e2728a.jpeg","article_res/cover/ba011d265e6d84b5c8cb6fd6b757b6cc.jpeg","Dall-E 3","DALL·E 3 understands significantly more nuance and detail, allowing you to easily translate your ideas into images.",[215,221,241],{"title":10,"list":216},[217,218,219,220],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":223},"GOOGLE",[224,225,226,234],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},615,"2025-03-30","#AI Researcher #AI Science #HKU #Google #AI Agent","article_res/cover/21fadf906067714bb0db31ae13a77c15.jpeg","article_res/cover/2697999a72bd26b22e85f0e92936d3ed.jpeg","AI-Researcher: LLM-driven全自动 scientific research assistant","AI-Researcher: Fully-Automated Scientific Discovery with LLM Agents  \nOpen-Sourced Alternative to Google AI Co-Scientist",{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},463,"2023-05-09","article_res/cover/89800f207723acdb55fc53bf999ebdc9.jpeg","article_res/cover/5764f369b4accd8f83e94aa4c077a175.jpeg","The Smallville sandbox world - A town with 25 virtual residents","Believable proxies of human behavior can empower interactive apps: Immersive environment, Rehearsal space, Prototyping tool",{"title":242,"list":243},"NVIDIA",[],true,{"code":4,"msg":5,"data":246},{"id":247,"publish_date":248,"is_original":23,"collection":249,"articles_id":250,"cover_url":251,"cover_url_1_1":252,"title":253,"summary":254,"author":28,"content":255,"popular":256,"list":320,"category":370,"tag":371},4,"2025-03-16","#VLM-R1 #Large Vision-Language Model #GRPO","aZf7ni-uEqUJkCEGKCBWUg","article_res/cover/262d831f5b3b6f3958d243586327d2f2.jpeg","article_res/cover/0b297b83de0f7b4572aac35ac4f14507.jpeg","VLM-R1: Enhancing Performance in Computer Vision Tasks Using GRPO","Solve Visual Understanding with Reinforced VLMs","\u003Cdiv class=\"rich_media_content js_underline_content\n                       autoTypeSetting24psection\n            \" id=\"js_content\">\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">VLM-R1 adopts the same GRPO algorithm as Deepseek R1, enhancing its visual capabilities. Today, let's explore how this algorithm improves performance in visual tasks.\u003C/span>\u003C/p>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100010356\" data-s=\"300,640\" data-type=\"png\" type=\"block\" style=\"height: auto !important;\" data-ratio=\"0.4564814814814815\" data-w=\"1080\" src=\"./assets/17423769061590.8342125036256682.png\">\u003C/section>\u003Ch3 style='box-sizing: border-box;margin: 30px 0px 15px;color: rgba(0, 0, 0, 0.85);font-weight: 500;cursor: pointer;padding: 0px;display: block;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;letter-spacing: normal;orphans: 2;text-align: left;text-indent: 0px;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan style=\"box-sizing: border-box;cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;text-align: left;font-weight: bold;display: block;\">\u003Cspan leaf=\"\">What is VLM-R1?\u003C/span>\u003C/span>\u003C/h3>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">GRPO (Group Relative Policy Optimization) helped Deepseek R1 improve its reasoning abilities; the VLM-R1 team found that GRPO can also help vision-language models (VLMs) perform better in general computer vision tasks, and its generalization ability surpasses traditional SFT (Supervised Fine-Tuning) methods.\u003C/span>\u003C/p>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">Currently, VLM-R1 performs exceptionally well, with a rapidly growing number of Stars.\u003C/span>\u003C/p>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100010345\" data-ratio=\"0.7212962962962963\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" type=\"block\" style=\"height: auto !important;\" src=\"./assets/17423769061730.36700563999033475.png\">\u003C/section>\u003Ch3 style='box-sizing: border-box;margin: 30px 0px 15px;color: rgba(0, 0, 0, 0.85);font-weight: 500;cursor: pointer;padding: 0px;display: block;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;letter-spacing: normal;orphans: 2;text-align: left;text-indent: 0px;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan style=\"box-sizing: border-box;cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;text-align: left;font-weight: bold;display: block;\">\u003Cspan leaf=\"\">Trial link\u003C/span>\u003C/span>\u003C/h3>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">https://huggingface.co/spaces/omlab/VLM-R1-Referral-Expression\u003C/span>\u003C/p>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100010346\" data-ratio=\"0.7222222222222222\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" type=\"block\" style=\"height: auto !important;\" src=\"./assets/17423769061620.9799909161204745.png\">\u003C/section>\u003Csection>\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection data-mpa-template=\"t\" mpa-from-tpl=\"t\" data-mpa-action-id=\"m7ljutsp1ip5\">\u003Csection data-mpa-template=\"t\" data-mpa-template-id=\"381\" data-mpa-category=\"模板\" mpa-from-tpl=\"t\">\u003Csection data-mpa-category=\"模板\" style=\"display: flex;justify-content: center;align-items: center;width: 100%;\" data-mid=\"\" mpa-from-tpl=\"t\">\u003Csection style=\"width: 96px;height: 10px;\" data-mid=\"\" mpa-from-tpl=\"t\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img\" data-ratio=\"0.10416666666666667\" data-w=\"192\" style=\"display: block;height: auto !important;\" src=\"./assets/17423769061550.550305849968487.png\">\u003C/section>\u003C/section>\u003C/section>\u003C/section>\u003Csection>\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100010349\" data-ratio=\"0.537962962962963\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" type=\"block\" style=\"height: auto !important;\" src=\"./assets/17423769077230.3829718707006813.png\">\u003C/section>\u003Csection mpa-from-tpl=\"t\" data-mpa-action-id=\"m7ljv0ex640\">\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection data-mpa-template=\"t\" mpa-from-tpl=\"t\" data-mpa-action-id=\"m7ljv0exyx\">\u003Csection data-mpa-template=\"t\" data-mpa-template-id=\"381\" data-mpa-category=\"模板\" mpa-from-tpl=\"t\">\u003Csection data-mpa-category=\"模板\" style=\"display: flex;justify-content: center;align-items: center;width: 100%;\" data-mid=\"\" mpa-from-tpl=\"t\">\u003Csection style=\"width: 96px;height: 10px;\" data-mid=\"\" mpa-from-tpl=\"t\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img\" data-ratio=\"0.10416666666666667\" data-w=\"192\" style=\"display: block;height: auto !important;\" src=\"./assets/17423769062160.7636722875419906.png\">\u003C/section>\u003C/section>\u003C/section>\u003C/section>\u003Csection>\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-ratio=\"0.637962962962963\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" type=\"block\" data-imgfileid=\"100010348\" style=\"height: auto !important;\" src=\"./assets/17423769065470.05414004811294859.png\">\u003C/section>\u003Csection mpa-from-tpl=\"t\" data-mpa-action-id=\"m7ljv2znjo1\">\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection data-mpa-template=\"t\" mpa-from-tpl=\"t\" data-mpa-action-id=\"m7ljv2znyey\">\u003Csection data-mpa-template=\"t\" data-mpa-template-id=\"381\" data-mpa-category=\"模板\" mpa-from-tpl=\"t\">\u003Csection data-mpa-category=\"模板\" style=\"display: flex;justify-content: center;align-items: center;width: 100%;\" data-mid=\"\" mpa-from-tpl=\"t\">\u003Csection style=\"width: 96px;height: 10px;\" data-mid=\"\" mpa-from-tpl=\"t\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img\" data-ratio=\"0.10416666666666667\" data-w=\"192\" style=\"display: block;height: auto !important;\" src=\"./assets/17423769062570.16856321866581792.png\">\u003C/section>\u003C/section>\u003C/section>\u003C/section>\u003Csection>\u003Cspan leaf=\"\">\u003Cbr>\u003C/span>\u003C/section>\u003Csection style=\"text-align: center;\" nodeleaf=\"\">\u003Cimg class=\"rich_pages wxw-img js_insertlocalimg\" data-imgfileid=\"100010347\" data-ratio=\"0.5453703703703704\" data-s=\"300,640\" data-type=\"png\" data-w=\"1080\" type=\"block\" style=\"height: auto !important;\" src=\"./assets/17423769063260.511672676893316.png\">\u003C/section>\u003Ch3 style='box-sizing: border-box;margin: 30px 0px 15px;color: rgba(0, 0, 0, 0.85);font-weight: 500;cursor: pointer;padding: 0px;display: block;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;letter-spacing: normal;orphans: 2;text-align: left;text-indent: 0px;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan style=\"box-sizing: border-box;cursor: pointer;font-size: 20px;color: rgb(0, 0, 0);line-height: 1.5em;letter-spacing: 0em;text-align: left;font-weight: bold;display: block;\">\u003Cspan leaf=\"\">Evaluation\u003C/span>\u003C/span>\u003C/h3>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">The team used the Qwen 2.5 VL 3B model for training on RefCOCO (a visual grounding task) and conducted evaluations on RefCOCO Val and RefGTA (an OOD task).\u003C/span>\u003C/p>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">Specifically, in the Referring Expression Comprehension (REC) task, the Qwen2.5-VL model was trained using the R1 and SFT methods. The results show that on in-domain test data, the performance of the SFT model is slightly lower than that of the R1 model.\u003C/span>\u003C/p>\u003Cp style='box-sizing: border-box;margin: 0px;cursor: pointer;color: rgb(0, 0, 0);font-size: 16px;line-height: 1.8em;letter-spacing: normal;text-align: left;text-indent: 0px;padding: 8px 0px;font-family: Optima, \"Microsoft YaHei\", PingFangSC-regular, serif;font-style: normal;font-variant-ligatures: normal;font-variant-caps: normal;font-weight: 400;orphans: 2;text-transform: none;widows: 2;word-spacing: 0px;-webkit-text-stroke-width: 0px;white-space: normal;background-color: rgb(255, 255, 255);text-decoration-thickness: initial;text-decoration-style: initial;text-decoration-color: initial;'>\u003Cspan leaf=\"\">However, on out-of-domain test data, the performance of the SFT model significantly decreases as the number of training steps increases, whereas the R1 model demonstrates stable improvement.\u003C/span>\u003C/p>\u003Cp style=\"display: none;\">\u003Cmp-style-type data-value=\"3\">\u003C/mp-style-type>\u003C/p>\u003C/div>",[257,265,274,282,289,297,305,313],{"id":258,"title_md5":259,"publish_date":260,"author_md5":261,"is_original":4,"collection":5,"summary_md5":262,"cover_url":263,"cover_url_1_1":264},537,"57e9d389d0af20e3d3b8296bf55b4c55","2022-05-19","cfab1ba8c67c7c838db98d666f02a132","50884c7903848f328c5bf87962e4e978","article_res/cover/22d7b30e0f60c7a1fe6c08aa70672dfc.jpeg","article_res/cover/200b2333171f80c88461cc35991c1685.jpeg",{"id":266,"title_md5":267,"publish_date":268,"author_md5":269,"is_original":23,"collection":270,"summary_md5":271,"cover_url":272,"cover_url_1_1":273},61,"28f6593a8424f2e07ba4205f545d9c52","2025-01-24","bc27fa490c4d0d525bac812fc0793534","#Research #LLM #AI Trader","1d06332ef50d8634094d0ab6ca4c213f","article_res/cover/07b5f567713288e88118d012b4cf9ef5.jpeg","article_res/cover/4679df1b91083fceb2013e8491ea61ae.jpeg",{"id":275,"title_md5":276,"publish_date":277,"author_md5":269,"is_original":4,"collection":278,"summary_md5":279,"cover_url":280,"cover_url_1_1":281},277,"3285acb478ecac410933c961a6fc84de","2024-05-12","#AI Index Report 2024 #Prompt Engineering #LLM","3ccba71c82af6187809edfcb1512a1a5","article_res/cover/6d277afc80e2551d0294bc3c6995b66a.jpeg","article_res/cover/c1ab171a5b7ca16284ee49b73ae116d2.jpeg",{"id":283,"title_md5":284,"publish_date":208,"author_md5":269,"is_original":4,"collection":285,"summary_md5":286,"cover_url":287,"cover_url_1_1":288},405,"1c169b30d542c329e065176a61c06a73","#AI Video Editor","f8ed5531f622d51c1ba6839889e2cd99","article_res/cover/fe640781e743aac4903acbe36ddda57e.jpeg","article_res/cover/f48646dafa6c7b086a8afebb640f3b3d.jpeg",{"id":290,"title_md5":291,"publish_date":292,"author_md5":293,"is_original":4,"collection":5,"summary_md5":294,"cover_url":295,"cover_url_1_1":296},530,"da032245150c1918d7303b320f288717","2022-05-26","8b3607d0f4181a3cb6ffdccf7185f09b","e2b45dcb478ccd0da9aaaac07a5d7bcc","article_res/cover/e6e7dfa9f1e4212e965aca0b1d570902.jpeg","article_res/cover/b0fa742d86b2b9f3815a0fd1e0cffb08.jpeg",{"id":298,"title_md5":299,"publish_date":300,"author_md5":269,"is_original":23,"collection":301,"summary_md5":302,"cover_url":303,"cover_url_1_1":304},65,"355241d059bad6536d0bb9bea822f270","2025-01-20","#AI Image Generator #AI Video Generator #Krea #AI 3D Generator","29bfdafcda8f22bb9177d697aa2e5446","article_res/cover/aa9dbc8db793695f3f24fcf41afc1fee.jpeg","article_res/cover/512cc57558a4f8a8eff6a2acbdb816a3.jpeg",{"id":306,"title_md5":307,"publish_date":308,"author_md5":269,"is_original":23,"collection":309,"summary_md5":310,"cover_url":311,"cover_url_1_1":312},183,"3f741220babab7d253633d147110226c","2024-09-03","#Meta","02d3cd60c5f2788d07872bf9c1e2a639","article_res/cover/1430550bbf1444488b94d767f1dd97c3.jpeg","article_res/cover/e06f0dabb08939da492aaaf263e5f07f.jpeg",{"id":314,"title_md5":315,"publish_date":316,"author_md5":269,"is_original":23,"collection":5,"summary_md5":317,"cover_url":318,"cover_url_1_1":319},255,"114ae9688a4d66229f2fb104c348995d","2024-06-05","3d251c4429a878c9491e67212b9f30f7","article_res/cover/5fa2e2c0b6838b32e2f9688ea2e0fbf5.jpeg","article_res/cover/8d6cb5e9fb78da2609b477ae5580e4a2.jpeg",{"related":321,"small":355},[322,329,337,346,347],{"id":323,"publish_date":324,"is_original":4,"collection":5,"cover_url":325,"cover_url_1_1":326,"title":327,"summary":328,"author":28},378,"2023-11-27","article_res/cover/a6e956c7192e834c23b9791f362717b6.jpeg","article_res/cover/fb444ef951eb25829e83c372cb080eb2.jpeg","Create GPTs using Zapier","Equip GPTs to take action in 3rd party apps with Zapier AI Actions!",{"id":330,"publish_date":331,"is_original":4,"collection":5,"cover_url":332,"cover_url_1_1":333,"title":334,"summary":335,"author":336},528,"2022-05-28","article_res/cover/056a477a27e0bb2f46e99b06d6e4588b.jpeg","article_res/cover/a1adf4fa5409d983ca82bd693eca9a3b.jpeg","Following Established Rules","Stop messing around","Excerpt",{"id":338,"publish_date":339,"is_original":4,"collection":340,"cover_url":341,"cover_url_1_1":342,"title":343,"summary":344,"author":345},155,"2024-10-10","#History of Intelligence #Neuroscience","article_res/cover/10e1e9838c22fb47444422912cf8e145.jpeg","article_res/cover/dcfca8f9c278e98010611cbd5e1d69df.jpeg","【A Brief History of Intelligence】6. Speaking Language (Human)","What makes us humans unique is that we possess a \"rational soul\" — the ability to reason, think abstractly, and reflect.  \n-Aristotle","Notes on \"A Brief History of Intelligence\"",{"id":174,"publish_date":175,"is_original":23,"collection":176,"cover_url":177,"cover_url_1_1":178,"title":179,"summary":180,"author":28},{"id":348,"publish_date":349,"is_original":23,"collection":350,"cover_url":351,"cover_url_1_1":352,"title":353,"summary":354,"author":28},5,"2025-03-15","#Meta #AI Video Generator","article_res/cover/6d289c936dcfa91f777ce3f553b01382.jpeg","article_res/cover/49d1007ac1b4cc08a3fa19c40455d915.jpeg","Meta's latest released video generation model VideoJAM","VideoJAM: Joint Appearance-Motion Representations for Enhanced Motion Generation in Video Models",[356,362,368],{"title":10,"list":357},[358,359,360,361],{"id":96,"publish_date":97,"is_original":23,"collection":98,"cover_url":99,"cover_url_1_1":100,"title":101,"summary":102,"author":28},{"id":104,"publish_date":105,"is_original":23,"collection":106,"cover_url":107,"cover_url_1_1":108,"title":109,"summary":110,"author":28},{"id":112,"publish_date":113,"is_original":23,"collection":114,"cover_url":115,"cover_url_1_1":116,"title":117,"summary":118,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"title":222,"list":363},[364,365,366,367],{"id":120,"publish_date":113,"is_original":23,"collection":121,"cover_url":122,"cover_url_1_1":123,"title":124,"summary":125,"author":28},{"id":166,"publish_date":167,"is_original":23,"collection":168,"cover_url":169,"cover_url_1_1":170,"title":171,"summary":172,"author":28},{"id":227,"publish_date":228,"is_original":23,"collection":229,"cover_url":230,"cover_url_1_1":231,"title":232,"summary":233,"author":28},{"id":235,"publish_date":236,"is_original":23,"collection":73,"cover_url":237,"cover_url_1_1":238,"title":239,"summary":240,"author":28},{"title":242,"list":369},[],[8,9,10],[8,12,13,14,9,10,15,16,17,18],["Reactive",245],1754646418957]