Understanding DeepSeek R1
Brodie Fell bu sayfayı düzenledi 3 ay önce


DeepSeek-R1 is an open-source language model developed on DeepSeek-V3-Base that's been making waves in the AI neighborhood. Not just does it match-or even surpass-OpenAI's o1 model in numerous benchmarks, however it likewise comes with totally MIT-licensed weights. This marks it as the very first non-OpenAI/Google design to provide strong reasoning abilities in an open and available manner.

What makes DeepSeek-R1 especially exciting is its openness. Unlike the less-open techniques from some market leaders, DeepSeek has published a detailed training method in their paper. The model is also remarkably affordable, with input tokens costing simply $0.14-0.55 per million (vs o1's $15) and output tokens at $2.19 per million (vs o1's $60).

Until ~ GPT-4, the typical knowledge was that better designs needed more information and calculate. While that's still valid, designs like o1 and R1 demonstrate an option: inference-time scaling through reasoning.

The Essentials

The DeepSeek-R1 paper provided several designs, but main among them were R1 and R1-Zero. Following these are a series of distilled designs that, while fascinating, I will not talk about here.

DeepSeek-R1 utilizes 2 major concepts:

1. A multi-stage pipeline where a little set of cold-start information kickstarts the design, wifidb.science followed by large-scale RL.

  1. Group Relative Policy Optimization (GRPO), a reinforcement learning method that counts on comparing multiple model outputs per timely to prevent the need for a different critic.

    R1 and R1-Zero are both reasoning models. This essentially implies they do Chain-of-Thought before addressing. For the R1 series of models, this takes kind as believing within a tag, before addressing with a final summary.

    R1-Zero vs R1

    R1-Zero uses Reinforcement Learning (RL) straight to DeepSeek-V3-Base with no supervised fine-tuning (SFT). RL is used to optimize the model's policy to take full advantage of reward. R1-Zero attains exceptional precision but often produces complicated outputs, bryggeriklubben.se such as blending several languages in a single response. R1 repairs that by integrating restricted monitored fine-tuning and numerous RL passes, which enhances both correctness and readability.

    It is interesting how some languages may express certain concepts better, which leads the model to choose the most meaningful language for the job.

    Training Pipeline

    The training pipeline that DeepSeek published in the R1 paper is tremendously fascinating. It showcases how they created such strong thinking models, and what you can expect from each stage. This consists of the issues that the resulting designs from each phase have, and how they fixed it in the next phase.

    It's interesting that their training pipeline varies from the normal:

    The typical training method: bbarlock.com Pretraining on large (train to predict next word) to get the base modelsupervised fine-tuningchoice tuning by means of RLHF R1-Zero: Pretrained → RL R1: Pretrained → Multistage training pipeline with multiple SFT and RL stages

    Cold-Start Fine-Tuning: Fine-tune DeepSeek-V3-Base on a couple of thousand Chain-of-Thought (CoT) samples to guarantee the RL process has a good beginning point. This provides a great model to start RL. First RL Stage: Apply GRPO with rule-based rewards to enhance thinking correctness and format (such as forcing chain-of-thought into thinking tags). When they were near merging in the RL procedure, they relocated to the next step. The outcome of this action is a strong reasoning model but with weak basic capabilities, [smfsimple.com](https://www.smfsimple.com/ultimateportaldemo/index.php?action=profile