
cerebras
cerebras/deepseek-r1-distill-llama-70b
DeepSeek R1 Distill Llama 70B model optimized for fast inference on Cerebras hardware. Supports up to 65,536 tokens context length.
Provider:
cerebras
Model type:
chat
Location:
US
Context Window
Intelligence Rating
Speed Rating
Cost Efficiency Rating
Pricing
$
Input tokens per million
$
Output tokens per million
