Blogs


When Size Isn’t Everything: Why Sapient’s 27M-Parameter HRM Matters for Small Models & AGI

What is HRM (and why we should care)

Singapore’s Sapient Intelligence introduced the Hierarchical Reasoning Model (HRM) — a 27M-parameter, brain-inspired, multi-timescale recurrent architecture trained with just 1,000 examples and no pre-training. According to the authors (arxiv.org), HRM outperforms GPT-o3-mini and Claude on the ARC-AGI benchmark, a test designed to measure genuine inductive reasoning rather than pattern replication.

The design mirrors cognitive neuroscience: the brain separates slow, global planning from fast, fine-grained execution. HRM encodes these separate timescales directly into its architecture.

Alt text

Empirical Results

Sapient reports:

  • ARC-AGI: HRM surpasses o3-mini-high, Claude 3.7 (8K), and DeepSeek R1 on Sapient’s internal ARC-AGI evaluations (coverage).
  • Structured reasoning tasks: Near-perfect results on Sudoku-Extreme and 30×30 Maze-Hard, where chain-of-thought-dependent LLMs typically break down.
  • Efficiency profile:
    • ~1,000 labeled examples
    • Zero pre-training
    • No chain-of-thought supervision
    • Single-pass inference
    • Over 90% reduction in compute relative to typical LLM reasoning pipelines (ACN Newswire)

The data suggests that architectural inductive bias can outperform sheer parameter scale.

Read on →

The $1.5 Trillion Question: Is AI Investment a Bubble or the Future?

The world is witnessing an investment phenomenon unlike anything since the dot-com boom. In 2024 alone, artificial intelligence companies attracted over $100 billion in venture capital funding, while semiconductor manufacturing has seen commitments exceeding $630 billion. Tech giants are pouring unprecedented sums into AI infrastructure, with some analysts now questioning whether this represents visionary transformation or dangerous overinvestment. The answer may determine the trajectory of the global economy for the next decade.

The Numbers Don’t Lie: A Historic Investment Surge

AI Funding Reaches Stratospheric Heights

The scale of AI investment in 2024-2025 defies historical precedent:

  • Global AI VC funding in 2024: $110 billion (up 80% from $55.6 billion in 2023)
  • Generative AI funding alone: $45 billion (nearly double 2023’s $24 billion)
  • 2025 trajectory: Through August, AI startups raised $118 billion, on pace to exceed 2024’s record
  • Market concentration: AI captured 33% of all global venture funding in 2024

To put this in perspective, AI investment in 2024 represented the highest funding year for the sector in the past decade, surpassing even the peak global funding levels of 2021. The late-stage deal sizes tell an even more dramatic story: average valuations jumped from $48 million in 2023 to $327 million in 2024 for generative AI companies.

Read on →

Attention Is All You Need: The Paper That Revolutionized AI

In June 2017, eight researchers from Google Brain and Google Research published a paper that would fundamentally reshape artificial intelligence. Titled “Attention Is All You Need,” it introduced the Transformer architecture—a model that discarded the conventional wisdom of sequence processing and replaced it with something elegantly simple: pure attention.

The numbers tell the story. As of 2025, this single paper has been cited over 173,000 times, making it one of the most influential works in machine learning history. Today, nearly every large language model you interact with—ChatGPT, Google Gemini, Claude, Meta’s Llama—traces its lineage directly back to this architecture.

But here’s what makes this achievement remarkable: it wasn’t about adding more layers, more parameters, or more complexity. It was about removing what had been considered essential for decades.

The Problem: Sequential Processing

Why RNNs Were Dominant (And Problematic)

Before 2017, the dominant approach for sequence tasks used Recurrent Neural Networks (RNNs), particularly Long Short-Term Memory (LSTM) networks and Gated Recurrent Units (GRUs). The idea was intuitive: process sequences one element at a time, maintaining a hidden state that captures information from previous steps.

Think of it like reading a book word by word, keeping a mental summary as you go.

The Fundamental Bottleneck: RNNs have an inherent constraint—they must process sequentially. The output at step t depends on the hidden state h_t, which depends on the previous state h{t-1}, which depends on h{t-2}, and so on. This creates an unbreakable chain.

From the paper:

“Recurrent models typically factor computation along the symbol positions of the input and output sequences. Aligning the positions to steps in computation time, they generate a sequence of hidden states h_t, as a function of the previous hidden state h_{t-1} and the input for position t.”

Read on →

From Sand to Stars: The Amazing Journey of Silicon Chips to Quantum Computing

Imagine if I told you that the most powerful computers in the world are made from the same stuff you find at the beach. You’d probably think I was kidding! But it’s absolutely true. Silicon, the second most common element in Earth’s crust, has been the secret ingredient powering every smartphone, laptop, and gaming console for over 50 years.

But here’s where the story gets really exciting: scientists have discovered that silicon has reached its limits, and they’re now building computers that work like magic tricks – welcome to the world of quantum computing!

Alt textSource: Internet

Before Silicon: The Stone Age of Computing

The Era of Vacuum Tubes (1940s-1950s)

Before silicon chips existed, computers were massive monsters that filled entire rooms. The first electronic computer, ENIAC, weighed 30 tons and used 17,468 vacuum tubes – think of old-fashioned light bulbs that glowed when electricity passed through them.

Mind-blowing fact: ENIAC consumed 150 kilowatts of power (enough to power 100 modern homes) and could perform 5,000 additions per second. Your smartphone today can perform over 1 billion operations per second while using less power than a single ENIAC vacuum tube!

Read on →

Generative AI in 2025: Global Trends, Breakthroughs and Future Horizons

Generative AI (GenAI) has transitioned from an experimental technology to a cornerstone of global innovation by 2025, reshaping industries, economies, and societal norms. This comprehensive overview draws on recent reports, surveys, and developments to explore the latest happenings in the GenAI space worldwide, while projecting likely future trajectories.

From surging investments and enterprise adoption to ethical dilemmas and regulatory frameworks, GenAI’s evolution reflects a blend of unprecedented potential and persistent challenges. We’ll examine key trends, regional variations, technological breakthroughs, and forward-looking predictions, incorporating data from authoritative sources like Stanford’s AI Index, McKinsey and Gartner.

In 2025, GenAI has seen explosive growth in enterprise adoption, particularly in functions like marketing, product development, and software engineering. Companies are investing heavily, with traffic surging 890% and budgets growing 60% through 2027. Breakthroughs include multimodal AI, where models process text, images, video, and audio, enabling applications in public sectors for better data search and citizen services. However, issues like AI-generated ransomware and deepfakes are rising, prompting global regulatory responses.

Alt textSource: Generated by Matplotlib

Read on →