Hi everyone! 👋
The AI conversation has moved beyond model size to architectural fit. As agentic systems enter production, treating LLMs as interchangeable is no longer viable.
LLMs are dividing into specialized architectures, each optimized for reasoning, perception, action, or efficiency. Agent performance now depends on matching the right architecture to the right role.
My aim with this newsletter is to break down the 10 core LLM architectures shaping modern AI agents and how to think about choosing the right one.
GPT (Generative Pretrained Transformer)
Let's start with the OG. GPT-style models are decoder-only transformers that revolutionized natural language processing. They work by predicting the next token given all previous tokens, simple in concept, powerful in execution.

How It Works:
Tokenization: Break text into manageable chunks
Massive pretraining: Learn patterns from billions of documents
Autoregressive generation: Predict one token at a time, using each prediction to inform the next
GPT architectures shine as generalist language models, developing an emergent understanding of context, intent, and language patterns through sheer exposure to human text. This makes them highly effective for natural conversations, ambiguous queries, coherent response generation, and code synthesis.
However, GPTs are fundamentally pattern predictors rather than reasoners. They struggle with multi-step logical thinking and can produce confident but incorrect outputs. They work well for conversational and content-focused agents, but once an agent needs to reason, plan, or verify decisions, GPT alone is no longer sufficient.
Best For: Conversational AI, content generation, code assistants, summarization, general-purpose language tasks
MoE (Mixture of Experts)
Here's where things get interesting. MoE architectures introduce sparse computation instead of running all parameters for every token; a gating network routes each token to a small subset of "expert" networks.

How It Works:
Gating Network: Routes tokens to specific experts
Expert Networks: Specialized feed-forward networks
Sparse Activation: Only top-k experts activated per token
Efficiency: Scale to billions of parameters without proportional compute
What makes MoE brilliant for agents? Cost efficiency at scale. You can build models with 100B+ parameters that only activate 10-20B per forward pass. This matters enormously when you're running thousands of agent interactions per hour. The trade-off? Training complexity and ensuring balanced expert utilization—but when done right, MoE delivers frontier performance at a fraction of the cost.
Best For: High-throughput production systems, multilingual models, cost-sensitive applications at scale
LRM (Large Reasoning Models)
LRMs represent a fundamental shift: these aren't just bigger language models, they're models explicitly trained for multi-step reasoning. Think of them as the difference between pattern matching and actual problem-solving.

How It Works:
Chain-of-Thought: Internal reasoning steps before final answer
Decomposition: Break complex problems into manageable sub-problems
Self-Verification: Models check their own reasoning
Training: Reinforcement learning on reasoning traces
The breakthrough here is training methodology. LRMs use reinforcement learning to reward correct reasoning paths, not just correct answers. This produces models that can tackle novel problems through systematic decomposition rather than memorization. For AI agents handling complex decision-making, financial analysis, medical diagnosis, and strategic planning, LRMs are becoming essential.
The latest models like OpenAI's o1 and DeepSeek-R1 showcase this paradigm: they spend more "time" thinking before responding, producing dramatically better results on tasks requiring genuine reasoning.
Best For: Complex problem-solving, mathematical reasoning, strategic planning, multi-step decision-making
VLM (Vision-Language Models)
VLMs bridge the gap between visual perception and language understanding. For agents operating in the real world or even just navigating GUIs, this multimodal capability is non-negotiable.

How It Works:
Dual Encoders: Separate image and text encoding pipelines
Multimodal Fusion: Joint attention mechanisms across modalities
Unified Output: Text generation conditioned on visual inputs
Training: Contrastive learning + caption generation
The power of VLMs becomes obvious when you need agents to understand screenshots, interpret diagrams, or navigate visual interfaces. Models like GPT-4V, Gemini, and Claude can now "see" and reason about images with remarkable accuracy. For building GUI automation agents or robots that need to understand their environment, VLMs are the foundation.
Best For: Visual question answering, GUI automation, accessibility tools, multimodal agents, document understanding
SLM (Small Language Models)
Not every problem needs a 70B parameter model. SLMs optimize for speed, cost, and deployability crucial when you're running agents on edge devices or need millisecond latency.

How It Works:
Distillation: Learning from larger teacher models
Architecture Optimization: Grouped-query attention, reduced dimensions
Quantization: Lower precision inference
Pruning: Remove redundant parameters
The real magic of SLMs isn't what they sacrifice, it's what they preserve. Through clever distillation and training techniques, models like Phi-3, Gemma, and Mistral 7B deliver surprisingly capable performance at 3-7B parameters. For agents that need to run locally, on mobile devices, or in resource-constrained environments, SLMs are the only viable option.
Best For: Edge deployment, mobile applications, real-time systems, cost-sensitive applications, on-device inference
LAM (Large Action Models)
LAMs represent the evolution from "understanding" to "doing." These models don't just generate text; they generate structured actions that get executed in the real world.

How It Works:
Action-Oriented Training: Fine-tuned on action trajectories
Structured Outputs: JSON, API calls, function parameters
Tool Integration: Native understanding of available tools
Feedback Loops: Learn from action outcomes
This is where things get exciting for practical AI agents. LAMs are trained specifically to understand the tools at their disposal and generate valid, executable actions. Rather than generating natural language that needs parsing, they produce structured outputs ready for execution. Think web navigation, API orchestration, robotic control, and LAMs are purpose-built for agents that need to do things, not just talk about them.
Best For: Web automation, API orchestration, task execution agents, robotic control, workflow automation
HRM (Hierarchical Reasoning Models)
HRMs solve a critical problem in agent architecture: how do you plan both strategically (high-level goals) and tactually (immediate actions) simultaneously?

How It Works:
Dual-Level Architecture: High-level planning + low-level execution
Hierarchical State Encoding: Abstract and concrete representations
Iterative Refinement: High-level plan guides low-level actions
Dynamic Adaptation: Adjust strategy based on execution feedback
The beauty of HRMs is their ability to maintain long-term coherence while remaining adaptive to immediate changes. The high-level planner thinks in terms of goals and strategies, while the low-level executor handles the nitty-gritty details. This hierarchical decomposition mirrors how humans actually plan complex tasks, and it's proving essential for agents tackling long-horizon problems.
Best For: Complex multi-step planning, long-horizon tasks, strategic decision-making, hierarchical workflows
mHC (Manifold-Constrained Hyper-Connections)
Now this is where architectural innovation gets really interesting. mHC isn't about what the model does; it's about how information flows through the model during training and inference.

How It Works:
Enhanced Connectivity: Multiple parallel residual streams instead of one
Manifold Constraints: Geometric constraints ensure training stability
Identity Mapping: Preserves gradient flow in very deep networks
Efficient Implementation: Only 6-7% training overhead
DeepSeek's mHC architecture solves a fundamental problem: as models get deeper, training becomes unstable. Traditional residual connections create a bottleneck; naive widening causes instability. mHC threads the needle by using manifold geometry to maintain stability while increasing the information capacity of inter-layer connections.
Why does this matter for agents? Because it enables more stable training of deeper, more capable models without the typical scaling headaches. DeepSeek demonstrated 2.1-2.3% performance improvements on reasoning tasks with their 27B parameter model—and crucially, the benefits scale with model size.
Best For: Very large models requiring training stability, long-context reasoning, scaling to frontier model sizes
World Models
Here's the paradigm shift that's been brewing: what if AI agents could mentally simulate the consequences of their actions before taking them? That's exactly what World Models enable.

How It Works:
Environment Modeling: Learn physics, causality, and dynamics
Predictive Simulation: Imagine future states given actions
Latent Representations: Compact encoding of world state
Action-Conditioned Prediction: Forecast outcomes of potential actions
World Models mark a fundamental shift away from pure language prediction. Instead of generating the next token, they model how the world itself evolves, making them essential for embodied agents, robotics, autonomous vehicles, and digital twins that must understand dynamics, causality, and physical constraints, not just labels and descriptions.
By enabling simulation, World Models allow agents to reason about “what if” scenarios, plan multi-step actions, and learn from imagined experience. As Yann LeCun notes, LLMs lack grounded world simulation and persistent internal state; World Models directly address this gap by providing the spatial and temporal reasoning modern autonomous systems require.
Best For: Robotics, autonomous vehicles, embodied AI, digital twins, simulation-based planning, physical task execution
LCM (Large Concept Models)
LCMs represent perhaps the most ambitious architectural direction: moving beyond token-by-token generation to reasoning over abstract concepts and high-level representations.

How It Works:
Concept Embeddings: Representations of abstract ideas beyond tokens
Hierarchical Reasoning: Operate at multiple levels of abstraction
Symbolic Integration: Bridge neural and symbolic reasoning
Compositional Generalization: Recombine concepts in novel ways
The core insight of LCMs is that token-level processing is too granular for true abstract reasoning. By operating on learned concept representations, these models can reason more efficiently about complex, abstract problems. Think of mathematical proofs, scientific reasoning, and strategic planning domains where high-level conceptual manipulation is more important than surface-level language generation.
While still emerging, LCMs point toward a future where models don't just process language but genuinely manipulate abstract ideas. For agents handling complex knowledge work, this architectural direction could be transformative.
Best For: Abstract reasoning, scientific discovery, mathematical proofs, knowledge synthesis, conceptual problem-solving
Practical Takeaway:
GPT – Generate answers and call tools as per input
MoE – Handle high-volume requests with efficient resource usage
LRM – Complex problem-solving with multi-step reasoning
VLM – Computer-using AI agents and visual task automation
SLM – On-device agents with limited computational resources
LAM – Autonomous task execution and workflow automation
HRM – Goal-driven agents needing internal planning without verbose outputs
LCM - Agents requiring high-level conceptual reasoning and coherence across languages and modalities
mHC - Long Horizon agents with better reasoning but higher cost
World Models - Still theoretical, but models that can navigate our real environment
What's Next: The 2025-2026 Trajectory
Models are improving at a rapid pace, and I would suggest you use the same strategy that most AI companies are using: “Don’t depend on a model - solve a use case or a problem.”
In the coming years, these models will keep on improving to a point where there will be a convergence in quality and ability. When we reach that point, we won’t need powerful models but rather a powerful strategy to solve a problem.
Thanks for reading.
— Rakesh


