AI-ML
Ollama v0.18.4
RESUMEN
What's Changed ggml: force flash attention off for grok by @rick-github in mlx: fix KV cache snapshot memory leak by @jessegross in * mlxrunner: schedule periodic snapshots during prefill by @jessegross
Descripción Detallada
What's Changed ggml: force flash attention off for grok by @rick-github in mlx: fix KV cache snapshot memory leak by @jessegross in mlxrunner: schedule periodic snapshots during prefill by @jessegross in doc: update vscode doc by @hoyyeva in Full Changelog:
Actualización de Ollama v0.18.4 con correcciones y mejoras.
- Desactivada la atención flash para grok.
- Corregido un problema de fuga de memoria en el caché KV.
- Programadas instantáneas periódicas durante el prellenado.
- Actualizada la documentación de VSCode.
A quién le importa
Todos los que usan Ollama.
Generado por IA · puede contener errores
Releases Relacionados
AI-ML
Ollama v0.23.0
## Claude Desktop with Ollama Launch <img width="1272" height="872" alt="ca1" src="https://github.com/user-attachments/assets/1d550e3f-0272-4429-8cb2-06d32344cb77" /> Claude Desktop is now supported with Ollama Launch. Both Claude Cowork and Claude Code are supported within the Claude Desktop Ap
AI-ML
Ollama v0.22.1
## What's Changed * Updated the **Gemma 4** renderer for thinking and tool calling improvements * Model recommendations are now updated without updating Ollama * Aligned the desktop app's launch page with `ollama launch` integrations * Fixed the Poolside integration title in `ollama launch`