GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models
Paper
• 2508.06471
• Published
• 206
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid
Mamba-Transformer Reasoning Model
Paper
• 2508.14444
• Published
• 43
Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality,
Long Context, and Next Generation Agentic Capabilities
Paper
• 2507.06261
• Published
• 67
MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning
Attention
Paper
• 2506.13585
• Published
• 273
Paper
• 2506.10910
• Published
• 66
Paper
• 2505.09388
• Published
• 335
MiMo: Unlocking the Reasoning Potential of Language Model -- From
Pretraining to Posttraining
Paper
• 2505.07608
• Published
• 82
Phi-4-reasoning Technical Report
Paper
• 2504.21318
• Published
• 54
Llama-Nemotron: Efficient Reasoning Models
Paper
• 2505.00949
• Published
• 41
CLIMB: CLustering-based Iterative Data Mixture Bootstrapping for
Language Model Pre-training
Paper
• 2504.13161
• Published
• 93
DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning
Paper
• 2504.07128
• Published
• 87
Rethinking Reflection in Pre-Training
Paper
• 2504.04022
• Published
• 80
OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training
Tokens
Paper
• 2504.07096
• Published
• 77
Paper
• 2503.19786
• Published
• 55
LIMO: Less is More for Reasoning
Paper
• 2502.03387
• Published
• 62
Skywork Open Reasoner 1 Technical Report
Paper
• 2505.22312
• Published
• 54
Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality
Data for Efficient and Accurate Code LLM
Paper
• 2503.17793
• Published
• 24
RedStone: Curating General, Code, Math, and QA Data for Large Language
Models
Paper
• 2412.03398
• Published
• 2
Nemotron-CC-Math: A 133 Billion-Token-Scale High Quality Math
Pretraining Dataset
Paper
• 2508.15096
• Published
• 5
RLBFF: Binary Flexible Feedback to bridge between Human Feedback &
Verifiable Rewards
Paper
• 2509.21319
• Published
• 8
StarCoder 2 and The Stack v2: The Next Generation
Paper
• 2402.19173
• Published
• 152
Nemotron-Cascade: Scaling Cascaded Reinforcement Learning for General-Purpose Reasoning Models
Paper
• 2512.13607
• Published
• 36
Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning
Paper
• 2512.20848
• Published
• 38
X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests
Paper
• 2601.06953
• Published
• 45