Hosted deepseek-r1-distill-qwen-32b
I made deepseek-r1-distill-qwen-32b
available as a hosted endpoint.
https://glama.ai/models/deepseek-r1-distill-qwen-32b
I couldn't find it with other providers. Maybe others will find it useful too.
This deployment is OpenAI compatible, and it is specifically optimized for handling long contexts and it is cheaper and faster than r1.
As far as I can tell based on the benchmarks, for codings tasks at least, this model outperforms DeepSeek-R1-Distill-Llama-70B
.
In the context of Cline,
- select "OpenAI Compatible" from model dropdown
- base URL
https://glama.ai/api/gateway/openai/v1
- enter your API key
- set model to
deepseek-r1-distill-qwen-32b