v0.11.7
📦 ollama
✨ 4 features🐛 4 fixes🔧 2 symbols
Summary
This release introduces the DeepSeek-V3.1 model and the preview of Turbo mode for running large models. Several bugs related to model loading, thinking tags, and tool call parsing have also been resolved.
Migration Steps
- To use Turbo mode via the CLI, create an account on ollama.com/signup and follow the documentation to authenticate your Ollama installation.
- When running DeepSeek-V3.1 via CLI with Turbo enabled, use: OLLAMA_HOST=ollama.com ollama run deepseek-v3.1
✨ New Features
- DeepSeek-V3.1 model is now available via Ollama.
- DeepSeek-V3.1 supports hybrid thinking, configurable via the 'think' parameter in the Ollama API.
- Thinking can be enabled/disabled in the Ollama CLI using /set think or /set nothink commands.
- Introduction of Turbo mode (in preview) to run large models like DeepSeek-V3.1 (671B parameters) using cloud hardware.
🐛 Bug Fixes
- Fixed an issue preventing multiple models from loading on CPU-only systems.
- Ollama now correctly handles models that skip outputting the initial <think> tag (e.g., DeepSeek-V3.1).
- Fixed an issue where text was emitted when a model did not output an opening <think> tag.
- Fixed incorrect parsing of tool calls that contained '{' or '}' characters.
🔧 Affected Symbols
DeepSeek-V3.1<think>