Skip to content

PaperCodex

Subscribe

Agentic Tool Use

MiniMax-M1: The First Open-Weight Hybrid-Attention Model for Long-Context Reasoning and Efficient AI Agents

MiniMax-M1: The First Open-Weight Hybrid-Attention Model for Long-Context Reasoning and Efficient AI Agents 3001

MiniMax-M1 is a breakthrough in open large language models: it’s the world’s first open-weight, large-scale hybrid-attention reasoning model. Designed for…

12/19/2025Agentic Tool Use, Long-context Reasoning, Software Engineering Agents
rStar2-Agent: A 14B Math Reasoning Model That Outsmarts 671B Models with Smarter, Tool-Aware Agentic Reasoning

rStar2-Agent: A 14B Math Reasoning Model That Outsmarts 671B Models with Smarter, Tool-Aware Agentic Reasoning 1356

In the rapidly evolving landscape of large language models (LLMs), bigger isn’t always better—smarter is. Enter rStar2-Agent, a 14-billion-parameter reasoning…

12/17/2025Agentic Tool Use, Mathematical Reasoning, Reinforcement Learning For Reasoning
Copyright © 2026 PaperCodex.
  • Facebook
  • YouTube
  • Twitter

PaperCodex