HuggingFace 每日AI论文速递

2025.01.14 | 数学推理提升,内存开销减少


Listen Later

本期的 11 篇论文如下:

[00:24] 📊 The Lessons of Developing Process Reward Models in Mathematical Reasoning(数学推理中过程奖励模型开发的经验教训)

[01:10] 🧠 Tensor Product Attention Is All You Need(张量积注意力机制是关键)

[01:53] 🤖 $\text{Transformer}^2$: Self-adaptive LLMs(Transformer²:自适应大型语言模型)

[02:34] 🎥 VideoAuteur: Towards Long Narrative Video Generation(视频导演:面向长篇叙事视频生成)

[03:22] 🌐 WebWalker: Benchmarking LLMs in Web Traversal(WebWalker:在网页遍历中评估大语言模型)

[04:08] 🩺 O1 Replication Journey -- Part 3: Inference-time Scaling for Medical Reasoning(O1复现之旅 -- 第三部分:医疗推理的推理时间扩展)

[04:50] 🗣 MinMo: A Multimodal Large Language Model for Seamless Voice Interaction(MinMo:一种用于无缝语音交互的多模态大型语言模型)

[05:41] 🔧 SPAM: Spike-Aware Adam with Momentum Reset for Stable LLM Training(SPAM:带动量重置的尖峰感知Adam优化器用于稳定LLM训练)

[06:25] 🩺 BIOMEDICA: An Open Biomedical Image-Caption Archive, Dataset, and Vision-Language Models Derived from Scientific Literature(BIOMEDICA:一个开放的生物医学图像-文本档案、数据集及从科学文献中衍生出的视觉语言模型)

[07:15] 🧪 ChemAgent: Self-updating Library in Large Language Models Improves Chemical Reasoning(ChemAgent:大型语言模型中自更新库提升化学推理能力)

[07:51] 🌐 UnCommon Objects in 3D(三维中的不常见物体)

【关注我们】

您还可以在以下平台找到我们,获得播客内容以外更多信息

小红书: AI速递

...more
View all episodesView all episodes
Download on the App Store

HuggingFace 每日AI论文速递By duan