Home: Remove old ollama config
This commit is contained in:
@ -123,6 +123,9 @@
|
||||
vscode
|
||||
# ghidra # launch with _JAVA_AWT_WM_NONREPARENTING=1 (use programs.ghidra)
|
||||
|
||||
# AI stuff
|
||||
# comfy-ui-cuda # Use module
|
||||
|
||||
# Unity Stuff
|
||||
# unityhub
|
||||
# rider-unity
|
||||
@ -182,35 +185,6 @@
|
||||
};
|
||||
|
||||
services = {
|
||||
ollama = {
|
||||
enable = false;
|
||||
acceleration = "cuda";
|
||||
# home = "/var/lib/ollama";
|
||||
|
||||
# loadModels = [
|
||||
# "deepseek-r1:8b" # Default
|
||||
# "deepseek-r1:14b"
|
||||
# ];
|
||||
|
||||
# https://github.com/ollama/ollama/blob/main/docs/faq.md#how-do-i-configure-ollama-server
|
||||
environmentVariables = {
|
||||
# Flash Attention is a feature of most modern models
|
||||
# that can significantly reduce memory usage as the context size grows.
|
||||
OLLAMA_FLASH_ATTENTION = "1";
|
||||
|
||||
# The K/V context cache can be quantized to significantly
|
||||
# reduce memory usage when Flash Attention is enabled.
|
||||
OLLAMA_KV_CACHE_TYPE = "q8_0"; # f16, q8_0 q4_0
|
||||
|
||||
# To improve Retrieval-Augmented Generation (RAG) performance, you should increase
|
||||
# the context length to 8192+ tokens in your Ollama model settings.
|
||||
OLLAMA_CONTEXT_LENGTH = "8192";
|
||||
};
|
||||
|
||||
host = "127.0.0.1";
|
||||
port = 11434;
|
||||
};
|
||||
|
||||
flatpak = {
|
||||
packages = [
|
||||
# "com.valvesoftware.Steam"
|
||||
|
||||
Reference in New Issue
Block a user