DeepSeek: R1 Distill Qwen 32B

Fiabilité 20%

deepseek/deepseek-r1-distill-qwen-32b

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense mode...

🧠 Intelligence & Données
Knowledge Cutoff: 2024-07-31
Tokenizer: Qwen
Moderation: ✅ Non
📅 Cycle de vie
Ajouté le: 29/01/2025
Spécifications
  • Provider & Modalité deepseek text->text
  • Fenêtre de contexte 32,768 tokens
  • Max Output Tokens 32,768
  • Support des Outils (Tools) Non supporté
🔍 Modèles similaires
Modèle Provider Input Output Contexte
Reka Edge
rekaai/reka-edge
rekaai $0.1000 $0.1000 16,384
LiquidAI: LFM2-24B-A2B
liquid/lfm-2-24b-a2b
liquid $0.0300 $0.1200 32,768
LiquidAI: LFM2.5-1.2B-Thinking (free)
liquid/lfm-2.5-1.2b-think...
liquid $0.0000 $0.0000 32,768
LiquidAI: LFM2.5-1.2B-Instruct (free)
liquid/lfm-2.5-1.2b-instr...
liquid $0.0000 $0.0000 32,768
Mistral: Mistral Small Creative
mistralai/mistral-small-c...
mistralai $0.1000 $0.3000 32,768