DeepSeek

DeepSeek R1 Zero

DeepSeek deepseek-r1-zero
Model Information
Slug deepseek-r1-zero
LLM.txt View
Release Date January 20, 2025
GPQA 0.733
Organization
Model Description
DeepSeek-R1-Zero is a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step. It's 671B parameters in size, with 37B active in an inference pass.

It demonstrates remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors.

DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. See [DeepSeek R1](/deepseek/deepseek-r1) for the SFT model.

Available at 2 Providers
Provider Type Model Name Original Model Input ($/1M) Output ($/1M) Free Actions
AIHubMix
AIHubMix
DeepSeek-R1-Zero
deepseek-ai/DeepSeek-R1-Zero $2.20 $2.20
Writingmate
Writingmate
Chat Code
DeepSeek: DeepSeek R1 Zero
deepseek/deepseek-r1-zero - -