Ollama: How to generate reproducible results

Posted . Visible to the public.

By default, Ollama produces non-deterministic output: The same prompt will yield slightly different results each time. To get reproducible outputs, you must set both temperature: 0 and a fixed seed in the request options.

curl http://localhost:11434/api/chat -d '{
  "model": "qwen3.5:35b",
  "messages": [
    {
      "role": "user",
      "content": "Extract the invoice total from this document",
      "images": ["iVBORw0KGg..."]
    },
  ],
  "options": {
    "seed": 101,
    "temperature": 0
  }
}'

That said, you might not really need reproducible outputs very often. If you chose to use an LLM in production, better get used to non-deterministic behavior.. :-)

Profile picture of Michael Leimstädtner
Michael Leimstädtner
Last edit
Michael Leimstädtner
License
Source code in this card is licensed under the MIT License.
Posted by Michael Leimstädtner to makandra dev (2026-04-16 12:32)