Ollama

notes:
Release list
0.13.3
RECENT
0.13.2
0.13.1
0.13.0
0.12.11
0.12.10
0.12.9
0.12.8
0.12.7
0.12.6
0.12.5
0.12.4
0.12.3
0.12.2
0.12.1
0.12.0
0.11.11
0.11.10
0.11.9
0.11.8
Release notes:

What's Changed

  • Flash attention is now enabled by default for vision models such as mistral-3, gemma3, qwen3-vl and more. This improves memory utilization and performance when providing images as input.
  • Fixed GPU detection on multi-GPU CUDA machines
  • Fixed issue where deepseek-v3.1 would always think even with thinking is disabled in Ollama's app

New Contributors

  • @chengcheng84 made their first contribution in https://github.com/ollama/ollama/pull/13265
  • @nathan-hook made their first contribution in https://github.com/ollama/ollama/pull/13256

Full Changelog: https://github.com/ollama/ollama/compare/v0.13.1...v0.13.2-rc0

Copyright © 2023 - All right reserved by Yadoc SAS