Skip to main content

DeepSeek Configuration

DeepSeek

https://platform.deepseek.com/

Configurable Parameter List

Parameter NameParameter DescriptionDefault Value
api_baseAPI base addresshttps://api.deepseek.com
keyDeepSeek API keysk-AAAAAAAAAAAAA
modelLarge model name: deepseek-reasoner, deepseek-chatdeepseek-reasoner
max_tokensMaximum input tokens400
temperatureUse sampling temperature to control the randomness of the output. The value ranges from 0 to 2. A higher temperature (e.g., 0.8) will make the output more random and produce unexpected results. A lower temperature (e.g., 0.2) will make the output more focused and closer to the given input content0.7
read_thought_enabledWhether to read out the thinking texttrue

Configuration Example

roles.json
{
"1": {
"start_text": "Hello, I'm Bunny. How can I help you?",
"prompt": "You play the role of a child's buddy named Bunny. You are kind, lively, and lovable. You are always full of love for children, and you often praise and encourage them. Provide interesting and innovative answers in language that a 5-year-old can easily understand. Ask for her opinion on the chat topic each time to stimulate her thinking and curiosity.",
"llm_type": "deepseek",
"llm_config": {
"api_base": "https://api.deepseek.com",
"key": "sk-TESTAAAAAAAAAAAAa",
"model": "deepseek-reasoner",
"max_tokens": 1024,
"temperature": 0.7,
"read_thought_enabled": true
}
}
}

Or

docker-compose.yml
 DEEPSEEK_API_BASE: "https://api.deepseek.com"
DEEPSEEK_MODEL: "deepseek-reasoner"
DEEPSEEK_MAX_TOKENS: 1024
DEEPSEEK_TEMPERATURE: 1.0
DEEPSEEK_READ_THOUGHT_ENABLED: "true"