1

Home: Remove old ollama config

This commit is contained in:
2026-03-19 17:32:15 +01:00
parent 9b43074248
commit 4cbe9885b2
4 changed files with 14 additions and 30 deletions

View File

@ -123,6 +123,9 @@
vscode
# ghidra # launch with _JAVA_AWT_WM_NONREPARENTING=1 (use programs.ghidra)
# AI stuff
# comfy-ui-cuda # Use module
# Unity Stuff
# unityhub
# rider-unity
@ -182,35 +185,6 @@
};
services = {
ollama = {
enable = false;
acceleration = "cuda";
# home = "/var/lib/ollama";
# loadModels = [
# "deepseek-r1:8b" # Default
# "deepseek-r1:14b"
# ];
# https://github.com/ollama/ollama/blob/main/docs/faq.md#how-do-i-configure-ollama-server
environmentVariables = {
# Flash Attention is a feature of most modern models
# that can significantly reduce memory usage as the context size grows.
OLLAMA_FLASH_ATTENTION = "1";
# The K/V context cache can be quantized to significantly
# reduce memory usage when Flash Attention is enabled.
OLLAMA_KV_CACHE_TYPE = "q8_0"; # f16, q8_0 q4_0
# To improve Retrieval-Augmented Generation (RAG) performance, you should increase
# the context length to 8192+ tokens in your Ollama model settings.
OLLAMA_CONTEXT_LENGTH = "8192";
};
host = "127.0.0.1";
port = 11434;
};
flatpak = {
packages = [
# "com.valvesoftware.Steam"