Windows下部署 DeepSeek‑R1 的可行路线
方案一 Ollama 快速部署与调用
1) 安装 Ollama Windows 版(安装后可用命令行执行 ollama 命令)
2) 启动服务(默认监听 11434 端口):
ollama serve
3) 拉取并运行模型:
ollama pull deepseek-r1:7b
ollama run deepseek-r1:7b
4) 验证 API(非流式):
curl http://localhost:11434/api/generate -d "{"prompt":"你好","stream":false}"
5) 如需局域网访问,设置环境变量:
OLLAMA_HOST=0.0.0.0
OLLAMA_PORT=11434
方案二 Transformers 直连与推理服务
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
pip install transformers>=4.30.0 accelerate
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_path = r"C:\models\deepseek_r1" # 本地模型目录
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
torch_dtype=torch.float16,
device_map="auto"
)
def generate(prompt, max_new_tokens=256):
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=max_new_tokens, do_sample=True, temperature=0.7)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
if __name__ == "__main__":
print(generate("用Python写一个快速排序。"))# pip install fastapi uvicorn
from fastapi import FastAPI
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch, uvicorn
app = FastAPI()
tokenizer = AutoTokenizer.from_pretrained(r"C:\models\deepseek_r1")
model = AutoModelForCausalLM.from_pretrained(
r"C:\models\deepseek_r1",
torch_dtype=torch.float16,
device_map="auto"
)
@app.post("/generate")
async def generate(prompt: str, max_new_tokens: int = 256):
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=max_new_tokens, do_sample=True, temperature=0.7)
return {"text": tokenizer.decode(outputs[0], skip_special_tokens=True)}
if __name__ == "__main__":
uvicorn.run(app, host="0.0.0.0", port=8000)方案三 可视化客户端连接本地模型
git clone https://github.com/deepseek-ai/cherry-studio.git
cd cherry-studio && pip install -r requirements.txt
python app.py --model-path C:modelsdeepseek_r1 --port 7860
URL: http://localhost:11434/api/generate
Model: deepseek-r1:7b
常见问题与排障要点