DeepSeek
Access high-performance chat and advanced reasoning models with competitive pricing and full support for the DeepSeek R1 thought process.
Table of contents
The DeepSeek provider offers high-performance chat and reasoning models with competitive pricing. NodeLLM supports both the DeepSeek-V3 chat model and the DeepSeek-R1 reasoning model.
Configuration
import { createLLM } from "@node-llm/core";
const llm = createLLM({
provider: "deepseek",
deepseekApiKey: process.env.DEEPSEEK_API_KEY // Optional if set in env
});
Specific Parameters
You can pass DeepSeek-specific parameters using .withParams().
const chat = llm.chat("deepseek-chat").withParams({
presence_penalty: 0.5,
frequency_penalty: 0.5,
top_p: 0.9
});
Features
- Models:
deepseek-chat: Optimized for speed and proficiency in broad tasks (DeepSeek-V3).deepseek-reasoner: Optimized for complex reasoning and problem solving (DeepSeek-R1).
- Tools: Supported on
deepseek-chat. - Reasoning: Access inner thought process text from
deepseek-reasoner. - Streaming: Full streaming support for all models.
- Structured Output: Supported via automated prompt engineering and
json_objectmode transitions.
Usage Details
DeepSeek provides OpenAI-compatible endpoints, but NodeLLM handles the specific capability differences (like reasoning vs tool support) automatically through its internal registry.