Hierarchical Reasoning Model

Aug 2025

  • Guan Wang
  • Jin Li
  • Yuhao Sun
  • Xing Chen
  • Changling Liu
  • Yue Wu
  • Meng Lu
  • Sen Song
  • Yasin Abbasi Yadkori

Abstract

Reasoning, the process of devising and executing complex goal-oriented action sequences, remains a critical challenge in AI. Current large language models (LLMs) primarily employ Chain-of-Thought (CoT) techniques, which suffer from brittle task decomposition, extensive data requirements, and high latency. Inspired by the hierarchical and multi-timescale processing in the human brain, we propose the Hierarchical Reasoning Model (HRM), a novel recurrent architecture that attains significant computational depth while maintaining both training stability and efficiency. HRM executes sequential reasoning tasks in a single forward pass without explicit supervision of the intermediate process, through two interdependent recurrent modules: a high-level module responsible for slow, abstract planning, and a low-level module handling rapid, detailed computations.

Introduction

Current large language models struggle with complex reasoning tasks, relying heavily on Chain-of-Thought (CoT) techniques that require extensive training data and suffer from brittle task decomposition. The Hierarchical Reasoning Model (HRM) addresses these limitations by drawing inspiration from the human brain's hierarchical and multi-timescale processing capabilities. With only 27 million parameters, HRM achieves exceptional performance on complex reasoning tasks using only 1000 training samples, operating without pre-training or CoT data.

Architecture Overview

HRM employs a novel recurrent architecture with two interdependent modules that work together to achieve significant computational depth while maintaining training stability.

High-Level Module

Responsible for slow, abstract planning and strategic decision-making. This module operates at a higher level of abstraction, focusing on long-term goals and overall task structure.

Low-Level Module

Handles rapid, detailed computations and immediate task execution. This module processes fine-grained details and implements the high-level plans through specific actions.

Single Forward Pass

Unlike traditional approaches that require multiple iterations, HRM executes sequential reasoning tasks in a single forward pass without explicit supervision of intermediate processes.

Key Innovations

HRM introduces several breakthrough innovations that address fundamental limitations in current reasoning systems.

Computational Efficiency

With only 27 million parameters, HRM achieves exceptional performance while maintaining training stability and efficiency, making it significantly more accessible than traditional large-scale models.

Minimal Training Data

The model operates effectively with only 1000 training samples, eliminating the need for extensive datasets that are typically required for complex reasoning tasks.

No Pre-training Required

HRM operates without pre-training or Chain-of-Thought data, making it more cost-effective and easier to deploy in various applications.

Performance Results

HRM demonstrates exceptional performance across a range of challenging reasoning tasks, outperforming much larger models with significantly longer context windows.

Complex Sudoku Puzzles

The model achieves nearly perfect performance on complex Sudoku puzzles, demonstrating its ability to handle structured reasoning tasks with multiple constraints.

Optimal Path Finding

HRM excels at finding optimal paths in large mazes, showcasing its capability for spatial reasoning and optimization problems.

Abstraction and Reasoning Corpus (ARC)

The model outperforms much larger models on ARC, a key benchmark for measuring artificial general intelligence capabilities, highlighting its superior reasoning abilities.

Applications and Impact

  • Universal computation and general-purpose reasoning systems
  • Efficient AI systems for resource-constrained environments
  • Scalable reasoning solutions without extensive training data
  • Real-time reasoning applications with low latency requirements

Future Directions

  • Scaling to even more complex reasoning tasks
  • Integration with existing AI systems and frameworks
  • Development of specialized variants for domain-specific reasoning
  • Exploration of multi-modal reasoning capabilities

Conclusion

The Hierarchical Reasoning Model represents a transformative advancement in AI reasoning capabilities. By drawing inspiration from human brain architecture and implementing efficient hierarchical processing, HRM achieves exceptional performance with minimal computational resources. This work underscores the potential for developing universal computation and general-purpose reasoning systems that are both powerful and accessible.

For the full details, see the original paper: Hierarchical Reasoning Model.