DiscoverSmart Enterprises: AI FrontiersMastering Reasoning LLMs: Decoding AI's Complex Problem-Solving Strategies
Mastering Reasoning LLMs: Decoding AI's Complex Problem-Solving Strategies

Mastering Reasoning LLMs: Decoding AI's Complex Problem-Solving Strategies

Update: 2025-07-29
Share

Description

Join us for an insightful exploration into the world of Reasoning LLMs, drawing on the expertise of Sebastian Raschka, PhD. This episode demystifies how Large Language Models (LLMs) are being refined to excel at complex tasks that require intermediate steps, such as solving puzzles, advanced mathematics, and challenging coding problems, moving beyond simple factual question-answering.

We'll uncover the four main approaches currently used to build and improve these specialised reasoning capabilities:

  • Inference-time scaling: Discover how techniques like Chain-of-Thought (CoT) prompting encourage LLMs to generate intermediate reasoning steps, mimicking a 'thought process' and often leading to more accurate results on more complex problems. This approach increases computational resources during inference, making it more expensive.
  • Pure Reinforcement Learning (RL): Learn about the surprising emergence of reasoning behaviour from pure reinforcement learning, as demonstrated by DeepSeek-R1-Zero. This model was trained exclusively with RL, without an initial supervised fine-tuning (SFT) stage, using accuracy and format rewards to develop basic reasoning skills.
  • Supervised Fine-tuning (SFT) + Reinforcement Learning (RL): Understand this key approach for building high-performance reasoning models, exemplified by DeepSeek's flagship R1 model. This method refines models with additional SFT stages and further RL training, building upon "cold-started" pure RL models.
  • Pure SFT and Distillation: Explore how smaller, more efficient reasoning models can be created by instruction fine-tuning them on high-quality SFT data generated by larger, stronger LLMs. This approach is particularly attractive for creating models that are cheaper to run and can operate on lower-end hardware.

We'll also discuss when to use reasoning models – they are ideal for complex challenges but can be inefficient, more verbose, and expensive for simpler tasks, sometimes even being "prone to errors due to 'overthinking'". The episode provides valuable insights from the DeepSeek R1 pipeline as a detailed case study and touches upon comparisons with models like OpenAI's o1. Plus, get tips for developing reasoning models on a limited budget, including the promise of distillation and innovative methods like 'journey learning', which includes incorrect solution paths to teach models from mistakes. Tune in to navigate the rapidly evolving landscape of reasoning LLMs!

Comments 
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Mastering Reasoning LLMs: Decoding AI's Complex Problem-Solving Strategies

Mastering Reasoning LLMs: Decoding AI's Complex Problem-Solving Strategies

Ali Mehedi