The issue is that this article is conflating (as do many, many articles about the topic) the distilled versions of R1 (basically llama/qwen reasoning finetunes) with the real thing. We are not even talking about quantized versions of R1 here, so it's not quite accurate to say you're running R1 here.
Hey the model from https://ollama.com/library/deepseek-r1 is used.
This is not the real thing? Just 671b is the real thing or what are you going to tell me?