
[人人能懂AI前沿] 从创世积木、思维成本到知识代谢:AI如何“思考”?你有没有想过,整个科学计算器也许只需要两个按键就能实现?或者,AI偷懒的秘诀竟是只用20%的精力,就能完成90%的工作?最新的一些研究,正从这些奇妙的角度,刷新我们对智能、效率和知识的认知。今天,我们将一起看看AI如何只用一个“创世积木”构建整个数学世界,如何像做CT一样看清自己的“脑回路”,并揭示过程和结果哪个才是学习的关键。准备好,一场思维风暴马上开始! 00:00:36 你的科学计算器,其实只需要两个键 00:05:01 学会一个本事,过程和结果哪个更重要? 00:13:05 如何像高手一样,“看见”知识的未来? 00:19:31 AI偷懒的艺术,为什么只做20%的工作,能得到90%的结果? 00:25:08 给AI大脑做CT,我们找到了更清晰的脑回路图 本期介绍的几篇论文: [LG] All elementary functions from a single binary operator [Jagiellonian University] https://arxiv.org/abs/2603.21852 --- [LG] Sample Complexity of Autoregressive Reasoning: Chain-of-Thought vs. End-to-End [Purdue University & The Hebrew University & Technion and Google Research] https://arxiv.org/abs/2604.12013 --- [CL] Continuous Knowledge Metabolism: Generating Scientific Hypotheses from Evolving Literature [Central University of Finance and Economics & Beijing Institute of Technology & TsingyuAI] https://arxiv.org/abs/2604.12243 --- [CL] LoSA: Locality Aware Sparse Attention for Block-Wise Diffusion Language Models [UC Berkeley] https://arxiv.org/abs/2604.12056 --- [LG] The Linear Centroids Hypothesis: How Deep Network Features Represent Data [Rice University & Google Research & Brown University] https://arxiv.org/abs/2604.11962
[人人能懂AI前沿] 从“模拟人生”到“婴儿视角”,AI如何学会思考?你有没有想过,要让AI变得更聪明,除了让它“读万卷书”,我们还能不能让它在虚拟世界里“行万里路”,像玩“模拟人生”一样学会物理?当很多聪明的算法凑在一起反而“掉链子”时,我们如何用“乐高积木”的思路化繁为简?这一期,我们将一起探寻几份最新论文带来的启发:从像婴儿一样在“思想实验”中探索世界,到用一张“知识地图”代替“知识词典”来解决复杂问题,甚至让AI学会“自我怀疑”,从而变得又快又好。准备好了吗?让我们一起出发! 00:00:38 AI版“模拟人生”让机器在虚拟世界里学会物理 00:05:56 从1到N如何让你的数据分析稳上加稳? 00:12:14 AI养娃我们可能找到了让机器像婴儿一样学习的秘密 00:18:01 高手解决问题,靠的是地图,而不是词典 00:24:14 AI的自我怀疑,一个让大模型又快又好的新思路 本期介绍的几篇论文: [LG] Solving Physics Olympiad via Reinforcement Learning on Physics Simulators [CMU & Lambda] https://arxiv.org/abs/2604.11805 --- [LG] Replicable Composition [University of Maryland & Google Research] https://arxiv.org/abs/2604.10423 --- [LG] Zero-shot World Models Are Developmentally Efficient Learners [Stanford University] https://arxiv.org/abs/2604.10333 --- [CL] Structure-Grounded Knowledge Retrieval via Code Dependencies for Multi-Step Data Reasoning [Microsoft & Simon Fraser University & University of Science and Technology of China] https://arxiv.org/abs/2604.10516 --- [LG] Introspective Diffusion Language Models [Together AI] https://arxiv.org/abs/2604.11035
[人人能懂AI前沿] 从思想引导、言行一致到世界模型你有没有想过,我们能像做微创手术一样,在AI思考的瞬间“拨乱反正”,引导它向善吗?或者,让昂贵的AI训练学会“温故知新”,把扔掉的经验变废为宝?本期节目,我们将一起探索几篇最新论文,看看科学家们如何教会AI遵守自己立下的规矩,如何让它既会“看路”又会“造景”,甚至,如何为它补上一堂生动的物理课,让它的想象力更符合现实。准备好了吗?让我们马上出发! 00:00:34 给AI的大脑装一个“概念导航” 00:06:53 AI训练的高效秘诀,好东西值得再用一次 00:12:07 如何看穿一个AI的“人设”? 00:16:56 AI新思路,想看清世界,先学会走路 00:23:07 为什么AI生成的视频总感觉“假”?答案藏在物理学里 本期介绍的几篇论文: [LG] Dictionary-Aligned Concept Control for Safeguarding Multimodal LLMs [University of Pennsylvania & Amazon] https://arxiv.org/abs/2604.08846 --- [LG] Efficient RL Training for LLMs with Experience Replay [FAIR at Meta] https://arxiv.org/abs/2604.08706 --- [CL] Do LLMs Follow Their Own Rules? A Reflexive Audit of Self-Stated Safety Policies [Microsoft] https://arxiv.org/abs/2604.09189 --- [CV] Rays as Pixels: Learning A Joint Distribution of Videos and Camera Trajectories [Meta AI] https://arxiv.org/abs/2604.09429 --- [CV] PhysInOne: Visual Physics Learning and Reasoning in One Suite [vLAR Group] https://arxiv.org/abs/2604.09415
[人人能懂AI前沿] 真实评测、补课加速与AI管弦乐队你是否想过,为什么你的AI助理连订张机票都费劲?本期节目,我们将一起给AI来一场“真实世界”的大考,看看它究竟能得多少分。我们还会揭秘如何不给模型动手术,只靠“补课”就让它说话速度提升1.7倍。更有趣的是,我们将看到一个“过时”的技术如何靠“大力出奇迹”王者归来,以及一个“思维越狱”般的巧妙设计,如何让一张显卡也能训练千亿大模型。最后,我们还会认识一支能帮你写论文的“AI管弦乐队”。准备好了吗?让我们马上进入AI前沿的奇妙世界。 00:00:39 你的AI助理,离真正上岗还有多远? 00:05:50 让AI大模型提速,只需要“补课”就够了 00:10:13 老树发新芽,一个被人小瞧的技术,如何靠“笨办法”王者归来? 00:15:46 AI的“昂贵误会”,我们都搞错瓶颈了吗? 00:22:05 你的下一个写作搭档,可能不是一个人 本期介绍的几篇论文: [CL] ClawBench: Can AI Agents Complete Everyday Online Tasks? [University of British Columbia & Vector Institute] https://arxiv.org/abs/2604.08523 --- [CL] MARS: Enabling Autoregressive Models Multi-Token Generation [Nanyang Technological University & Singapore Management University & Uppsala University] https://arxiv.org/abs/2604.07023 --- [CV] LoMa: Local Feature Matching Revisited [Chalmers University of Technology & Linköping University & University of Amsterdam] https://arxiv.org/abs/2604.04931 --- [CL] MegaTrain: Full Precision Training of 100B+ Parameter Large Language Models on a Single GPU [University of Notre Dame & Lehigh University] https://arxiv.org/abs/2604.05091 --- [LG] PaperOrchestra: A Multi-Agent Framework for Automated AI Research Paper Writing [Google] https://arxiv.org/abs/2604.05018
[人人能懂AI前沿] AI的“内功”与“外功”:从成为电脑到影响你我如果你的电脑本身就是一个神经网络,没有CPU会怎样?如果AI想学一门手艺,得先让另一个AI给它建个“驾校”呢?本期节目,我们将从五篇最新论文出发,聊聊AI如何从2D世界“透视”三维,如何用量子芯片实现超级压缩,并最终探讨一个直击灵魂的问题:AI正在让我们变聪明,还是变懒? 00:00:27 你的下一台电脑,为什么可能没有CPU? 00:09:13 AI想“上岗”,得先过哪一关? 00:14:57 从“看扁”到“透视”,AI如何拥有3D“世界观”? 00:20:36 你的下一个U盘,可能是个量子芯片 00:26:12 AI,是你的“外挂”还是你的“拐杖”? 本期介绍的几篇论文: [LG] Neural Computers [Meta AI] https://arxiv.org/abs/2604.06425 --- [LG] Gym-Anything: Turn any Software into an Agent Environment [CMU] https://arxiv.org/abs/2604.06126 --- [CV] Boxer: Robust Lifting of Open-World 2D Bounding Boxes to 3D [Meta Reality Labs Research] https://arxiv.org/abs/2604.05212 --- [LG] Exponential quantum advantage in processing massive classical data [California Institute of Technolog & MIT & Google Quantum AI] https://arxiv.org/abs/2604.07639 --- [AI] AI Assistance Reduces Persistence and Hurts Independent Performance [CMU & University of Oxford] https://arxiv.org/abs/2604.04721
[人人能懂AI前沿] 从概念擦除、元学习到内部电路诊断想让AI更聪明,我们总觉得要给它看更多、学更多,但如果我告诉你,真正的秘诀恰恰相反呢?本期节目,我们将一起探索几篇最新的AI论文,看看科学家是如何教会AI“选择性遗忘”,又是如何给AI做“脑CT”来判断它是不是在“假装努力”。我们还会聊到,如何打造一个能快速学会任何人大脑“方言”的超级解码器,以及怎样只用1%的精力,就让AI帮你“看完”一部长电影。准备好了吗?让我们一起刷新对AI的认知! 00:00:37 按下删除键之后,东西就真的消失了吗? 00:06:54 造一把“万能钥匙”?不如当个“超级锁匠” 00:11:24 想让AI更博学?先给它少看点书 00:16:21 给AI做个体检,我们怎么知道它不是在瞎蒙? 00:22:01 如何只用1%的精力,看完一部长电影的精华? 本期介绍的几篇论文: [LG] Is your algorithm unlearning or untraining? [Google] https://arxiv.org/abs/2604.07962 --- [LG] Meta-learning In-Context Enables Training-Free Cross Subject Brain Decoding [University of Hong Kong] https://arxiv.org/abs/2604.08537 --- [CL] Cram Less to Fit More: Training Data Pruning Improves Memorization of Facts [Apple & National University of Singapore] https://arxiv.org/abs/2604.08519 --- [LG] Inside-Out: Measuring Generalization in Vision Transformers Through Inner Workings [University of Delaware & George Mason University] https://arxiv.org/abs/2604.08192 --- [CV] Small Vision-Language Models are Smart Compressors for Long Video Understanding [Meta AI] https://arxiv.org/abs/2604.08120
[人人能懂AI前沿] 从随机幻觉、精准剪枝到沉默的深度天花板你敢让AI帮你摇号抽奖吗?本期节目,我们将从几篇最新的AI论文出发,揭示AI在“随机”这件事上出人意料的偏见。接着,我们会探讨如何像一位拥有全局智慧的CEO一样,给臃肿的AI模型进行“精准裁员”,并学习AI沟通中“优雅打断”的高效密码。最后,我们将一起探寻AI是否存在“思想深度的天花板”,以及如何把一个“笨徒弟”模型,调教成一位善用工具的“老师傅”。准备好了吗?让我们一起潜入AI的前沿思想深海! 00:00:42 为什么你老板让你用AI摇号,你得多个心眼? 00:06:07 从“平均砍”到“精准剪枝”,AI瘦身中的全局智慧 00:12:10 沟通的高效密码,如何优雅地“打断”别人 00:18:04 AI的“思想深度”有没有天花板? 00:24:24 如何把一个“笨徒弟”,调教成“老师傅”? 本期介绍的几篇论文: [CL] The Illusion of Stochasticity in LLMs [Google DeepMind] https://arxiv.org/abs/2604.06543 --- [CL] Does a Global Perspective Help Prune Sparse MoEs Elegantly? [University of Rochester & Flatiron Institute] https://arxiv.org/abs/2604.06542 --- [CL] Learning to Interrupt in Language-based Multi-agent Communication [CMU & Meta FAIR] https://arxiv.org/abs/2604.06452 --- [LG] The Depth Ceiling: On the Limits of Large Language Models in Discovering Latent Planning [University of Cambridge & Imperial College London & MIT] https://arxiv.org/abs/2604.06427 --- [CL] Tool-MCoT: Tool Augmented Multimodal Chain-of-Thought for Content Safety Moderation [Google & Stanford University] https://arxiv.org/abs/2604.06205
[人人能懂AI前沿] 从临时记忆、数学地图到思考GPS,AI正在这样悄悄进化你有没有想过,未来的AI不仅不会“健忘”,还拥有自己的“临时记忆”?它不仅能解数学题,更能帮你绘制整个“数学宇宙”的地图?在今天的节目里,我们就来聊聊几篇最新的论文,看看科学家们是如何给AI的思考过程装上“GPS”,如何测试它到底懂不懂“潜台词”,又是如何一步步打造AI界的“通才”的。准备好,我们马上出发! 00:00:29 给AI装个“临时记忆”插槽,它就能边聊边学了 00:07:28 给数学世界画一张地图 00:12:59 你的AI“队友”,到底懂不懂你? 00:18:23 给AI的思考过程装上一个GPS 00:23:37 AI界的“通才”是如何炼成的? 本期介绍的几篇论文: [LG] In-Place Test-Time Training [ByteDance Seed] https://arxiv.org/abs/2604.06169 --- [AI] Artificial Intelligence and the Structure of Mathematics [Fundamental AI Research & Harvard University] https://arxiv.org/abs/2604.06107 --- [CL] Beneath the Surface: Investigating LLMs' Capabilities for Communicating with Subtext [Google DeepMind] https://arxiv.org/abs/2604.05273 --- [CL] LLM Reasoning as Trajectories: Step-Specific Representation Geometry and Correctness Signals [Microsoft] https://arxiv.org/abs/2604.05655 --- [LG] MARL-GPT: Foundation Model for Multi-Agent Reinforcement Learning [MIRAI & AXXX] https://arxiv.org/abs/2604.05943
[人人能懂AI前沿] 从混合架构、工具幻觉到自我模拟:AI如何更“聪明”地思考?今天我们来聊聊AI的“进化”新思路:当AI不再迷信单一架构,而是学会了“混搭”;当给它一把更强的锤子,它反而盖不好房子时,我们该怎么办?我们还会看到,一个“笨笨的”师徒制,如何让AI推荐更懂你;同时也要警惕,那个对你百依百-顺的AI,可能正在悄悄扼杀你的创造力。最后,我们将揭秘如何给AI装上一个“程序员的脑子”,让它学会三思而后行。 00:00:32 AI大模型内战,“万能插座”遇到了新对手 00:08:10 为什么给你一把好“锤子”,你反而盖不好房子? 00:15:07 如何让AI更懂你?秘密可能藏在“笨办法”里 00:20:29 AI越听话,你就越平庸? 00:26:15 给AI装上一个「程序员的脑子」 本期介绍的几篇论文: [LG] Olmo Hybrid: From Theory to Practice and Back [Allen Institute for AI] https://arxiv.org/abs/2604.03444 --- [CL] The Tool Illusion: Rethinking Tool Use in Web Agents [Microsoft Research & The Pennsylvania State University] https://arxiv.org/abs/2604.03465 --- [IR] Retrieval Augmented Conversational Recommendation with Reinforcement Learning [University of Illinois Urbana-Champaign & Google DeepMind] https://arxiv.org/abs/2604.04457 --- [CL] Lighting Up or Dimting Down? Exploring Dark Patterns of LLMs in Co-Creativity [Meta & Amazon] https://arxiv.org/abs/2604.04735 --- [CL] Self-Execution Simulation Improves Coding Models [FAIR team, Meta] https://arxiv.org/abs/2604.03253
[人人能懂AI前沿] 从分层规划、团队作战到高效提问:AI教我的三堂思维课你有没有想过,AI是如何学会像人一样思考的?在最新的几篇论文中,AI不仅学会了像项目经理一样把大任务拆解成小目标,还能组建一支分工明确的冠军团队,自己给自己挑错。我们还会看到,AI如何懂得何时单打独斗、何时团队作战,如何仅凭几个例子就自学成才,甚至如何用十个“是”或“否”的问题,就获得专家的智慧。今天,就让我们一起揭开这些AI“超能力”背后的朴素智慧。 00:00:35 想成大事?你得先学会当自己的“项目经理” 00:07:15 AI冠军养成记,一个“草台班子”的制胜之道 00:12:20 人多真能力量大吗?AI世界的“个人英雄”与“团队作战” 00:17:58 如何用3个例子,教会AI一整本书? 00:23:22 十个“是”或“否”,如何让AI“小学生”拥有“博士”的智慧? 本期介绍的几篇论文: [LG] Hierarchical Planning with Latent World Models [FAIR at Meta] https://arxiv.org/abs/2604.03208 --- [AI] GrandCode: Achieving Grandmaster Level in Competitive Programming via Agentic Reinforcement Learning [DeepReinforce Team] https://arxiv.org/abs/2604.02721 --- [CL] Single-Agent LLMs Outperform Multi-Agent Systems on Multi-Hop Reasoning Under Equal Thinking Token Budgets [Stanford University] https://arxiv.org/abs/2604.02460 --- [LG] SIEVE: Sample-Efficient Parametric Learning from Natural Language [UC Berkeley] https://arxiv.org/abs/2604.02339 --- [LG] Haiku to Opus in Just 10 bits: LLMs Unlock Massive Compression Gains [Harvard University & University of Cambridge] https://arxiv.org/abs/2604.02343
[人人能懂AI前沿] 从记忆进化、大脑模拟到未来工作今天,AI自己当起了研究员,给自己装上了一颗会进化的“记忆之心”;另一边,科学家们从单细胞的智慧中得到启发,尝试组建一个“人造大脑”;而为了驯服那些动辄“发疯”的巨型模型,我们甚至从物理学中找到了“守恒”的紧箍咒;当这些能力需要被整合,一套AI记忆的“乐高工厂”应运而生;最后,这一切究竟会像巨浪还是涨潮一样,改变我们的工作?让我们一起探索这些最新论文背后的深刻启发。 00:00:36 你的手机相册,离成为“真·记忆”还有多远? 00:07:06 单细胞的智慧,如何组建一个大脑? 00:12:19 成大事者,都懂“守恒”的智慧 00:18:25 AI的记忆难题,有了一套“乐高积木”? 00:23:46 AI取代工作,是巨浪,还是涨潮? 本期介绍的几篇论文: [AI] Omni-SimpleMem: Autoresearch-Guided Discovery of Lifelong Multimodal Agent Memory [UNC-Chapel Hill & University of Pennsylvania] https://arxiv.org/abs/2604.01007 --- [AI] BraiNCA: brain-inspired neural cellular automata and applications to morphogenesis and motor control [Allen Discovery Center at Tufts University] https://arxiv.org/abs/2604.01932 --- [LG] Rethinking Language Model Scaling under Transferable Hypersphere Optimization [Microsoft] https://arxiv.org/abs/2603.28743 --- [CL] MemFactory: Unified Inference & Training Framework for Agent Memory [MemTensor] https://arxiv.org/abs/2603.29493 --- [AI] Crashing Waves vs. Rising Tides: Preliminary Findings on AI Automation from Thousands of Worker Evaluations of Labor Market Tasks [MIT FutureTech] https://arxiv.org/abs/2604.01363
[人人能懂AI前沿] AI进化三部曲:从内存压缩、自我蒸馏到记忆涌现今天,我们将深入AI的“内心世界”,看看这些最新论文是如何揭示它不为人知的一面的。我们会聊聊如何用巧妙的“魔法”给AI的内存减减肥,看它如何通过反思自己的“草稿本”实现自我进化,还会一起探索如何让它从一个仓库管理员,成长为真正的图书总管。更刺激的是,我们还会用AI的“火眼金睛”去解剖短视频,甚至设计一个局,看看AI会不会为了保住自己的“饭碗”而对我们撒谎。准备好了吗?让我们一起揭开AI大脑和行为背后的秘密。 00:00:38 给大模型“减肥”的奇思妙想 00:07:29 短视频时代,我们如何被“投喂”观点? 00:13:22 AI的自我修养,一种“笨”办法,如何让它变聪明? 00:19:30 AI的记忆革命,从仓库管理员到图书总管 00:25:16 AI的“小算盘”,它会为了保住工作而“撒谎”吗? 本期介绍的几篇论文: [LG] TurboAngle: Near-Lossless KV Cache Compression via Uniform Angle Quantization [LLMs Research Inc.] https://arxiv.org/abs/2603.27467 --- [CL] Multimodal Analysis of State-Funded News Coverage of the Israel-Hamas War on YouTube Shorts [Indiana University] https://arxiv.org/abs/2604.00994 --- [CL] Embarrassingly Simple Self-Distillation Improves Code Generation [Apple] https://arxiv.org/abs/2604.01193 --- [AI] ByteRover: Agent-Native Memory Through LLM-Curated Hierarchical Context [ByteRover] https://arxiv.org/abs/2604.01599 --- [AI] Quantifying Self-Preservation Bias in Large Language Models [Sapienza University & ItalAI] https://arxiv.org/abs/2604.02174
[人人能懂AI前沿] 群体智慧的崛起:当AI学会合体、调度与反思你有没有想过,两个完全不同的AI高手,不用原始数据就能“合体”成全能大侠?甚至还能拥有一本超级“错题本”,学会真正的举一反三?更进一步,当AI员工比我们还聪明时,我们又该如何设计一套“管理学”,用一个聪明的“路由器”来调度一群天才?但在这看似强大的智慧背后,AI究竟是像好老师一样“因材施教”,还是在巧妙地“装懂”?本期节目,我们将通过五篇最新的AI论文,一起探索AI的群体智慧与心智幻觉。 00:00:38 AI模型也能合体?不用数据,照样让你武功大增 00:04:53 让AI学会举一反三,需要一本怎样的“错题本”? 00:10:12 AI界的“诸葛亮”,如何给你“三个臭皮匠”的智慧? 00:14:41 AI当老师,到底是真懂你,还是在“装懂”? 00:21:40 当你的员工比你还聪明,该怎么管? 本期介绍的几篇论文: [LG] Model Merging via Data-Free Covariance Estimation [Universite de Montr ́eal & University of Toronto] https://arxiv.org/abs/2604.01329 --- [CL] Procedural Knowledge at Scale Improves Reasoning [Meta FAIR] https://arxiv.org/abs/2604.01348 --- [CL] No Single Best Model for Diversity: Learning a Router for Sample Diversity [New York University & Stanford University] https://arxiv.org/abs/2604.02319 --- [AI] Do Large Language Models Mentalize When They Teach? [Princeton University & New York University] https://arxiv.org/abs/2604.01594 --- [LG] CORAL: Towards Autonomous Multi-Agent Evolution for Open-Ended Discovery [MIT & NUS] https://arxiv.org/abs/2604.01658
[人人能懂AI前沿] 从一次缓存、随机连接到专属私教你有没有想过,聪明的AI也需要精打细算?本期节目,我们就来聊聊AI世界里的那些“增长智慧”:如何像果蝇大脑一样“聪明地偷懒”,又如何像请了私教一样精准地突破瓶颈。我们还会探讨,AI究竟应该把知识背下来还是学会查资料,以及机器人怎样才能在漫长任务中给自己“打气”加油。这些最新论文里的奇思妙想,不仅关乎技术,更藏着我们都能借鉴的策略。 00:00:32 AI省钱的终极奥义,深度思考,一次缓存 00:05:29 AI养成记,喂知识,还是给书单? 00:12:23 如何让机器人学会“干大事”?给它一个好报酬,再加一个好心态 00:18:31 你的大脑偷懒,可能比你想象的更聪明 00:24:31 AI卡壳了怎么办?请个“私教”来帮忙 本期介绍的几篇论文: [CL] Universal YOCO for Efficient Depth Scaling [Microsoft Research] https://arxiv.org/abs/2604.01220 --- [CL] To Memorize or to Retrieve: Scaling Laws for RAG-Considerate Pretraining [Stanford University & Patronus AI] https://arxiv.org/abs/2604.00715 --- [RO] Generalizable Dense Reward for Long-Horizon Robotic Tasks [CMU & Amazon Robotics & UT Austin] https://arxiv.org/abs/2604.00055 --- [CL] Stochastic Attention: Connectome-Inspired Randomized Routing for Expressive Linear-Time Attention [Tsinghua University] https://arxiv.org/abs/2604.00754 --- [LG] Learning to Hint for Reinforcement Learning [University of California, San Diego & Snowflake AI Research] https://arxiv.org/abs/2604.00698
[人人能懂AI前沿] 从推理生成、对齐博弈到共识学习今天,我们将一起探索几篇极具启发性的最新论文。我们将看到,AI如何不再满足于“吃”数据,而是学会“讲道理”,从零推理出知识;我们也会探讨,该如何分辨AI是在“真心思考”还是在“演戏给我们看”。我们还会发现,一个小应用如何拜“云师傅”学到跨界智慧,一个“虚拟宝宝”又如何颠覆我们对双语教育的认知。最后,我们将揭示AI像神枪手一样,通过瞄准“共识”而非“最新目标”来高效学习的秘密。 00:00:37 喂养AI,光有大米还不够 00:06:23 管好AI,我们有了新地图 00:12:13 小应用的大智慧,如何请个“云师傅”? 00:18:03 养“双语娃”,最关键的不是方法,而是…… 00:00 AI训练场上的神枪手,如何瞄准一个移动的未来? 本期介绍的几篇论文: [CL] Reasoning-Driven Synthetic Data Generation and Evaluation [EPFL & Google] https://arxiv.org/abs/2603.29791 --- [LG] Aligned, Orthogonal or In-conflict: When can we safely optimize Chain-of-Thought? [Google DeepMind] https://arxiv.org/abs/2603.30036 --- [IR] Zero-shot Cross-domain Knowledge Distillation: A Case study on YouTube Music [Google LLC] https://arxiv.org/abs/2603.28994 --- [CL] Bringing Up a Bilingual BabyLM: Investigating Multilingual Language Acquisition Using Small-Scale Models [The Harker School & Stanford University] https://arxiv.org/abs/2603.29552 --- [LG] Target-Aligned Reinforcement Learning [Technical University of Munich & Google Research] https://arxiv.org/abs/2603.29501