Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/index.php?threads/understanding-reasoning-llms.22038/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021770
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

Understanding Reasoning LLMs

XYang2023

Well-known member
This article describes the four main approaches to building reasoning models, or how we can enhance LLMs with reasoning capabilities. I hope this provides valuable insights and helps you navigate the rapidly evolving literature and hype surrounding this topic.

In 2024, the LLM field saw increasing specialization. Beyond pre-training and fine-tuning, we witnessed the rise of specialized applications, from RAGs to code assistants. I expect this trend to accelerate in 2025, with an even greater emphasis on domain- and application-specific optimizations (i.e., "specializations").

Variants of Deepseek R1 models:

1738874664493.png


 
From the technical article:

"The DeepSeek R1 technical report states that its models do not use inference-time scaling."

Relevant news:


"Looking ahead, the chip giant remains confident that the demand for AI compute will continue to grow steadily, driven by factors such as model pre-training, post-training, synthetic data generation, and new models like OpenAI's offerings that enable test-time scaling for inference."
 
Back
Top