diff --git a/.env.example b/.env.example index 71e09c3..1973dba 100644 --- a/.env.example +++ b/.env.example @@ -25,8 +25,6 @@ OLLAMA_DATA=/home/${STACK_USER}/.ollama # ── Ollama ──────────────────────────────────────────────────────────────────── OLLAMA_PORT=11434 -OLLAMA_SSH_KEY=/home/${STACK_USER}/.ssh/id_ed25519_ollama -OLLAMA_SSH_KEY_PUB=/home/${STACK_USER}/.ssh/id_ed25519_ollama.pub # Memory limits — adjust for your hardware. # Intel Arc users: 16gb / 32g is a good starting point for a 32GB system. # NVIDIA GPU users: match shm to your VRAM size. diff --git a/docker-compose.yml b/docker-compose.yml index b91df65..ff5507d 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -48,8 +48,6 @@ services: - OLLAMA_KEEP_ALIVE=-1 volumes: - ${OLLAMA_DATA:-/home/user/.ollama}:/root/.ollama - - ${OLLAMA_SSH_KEY:-/dev/null}:/root/.ollama/id_ed25519:ro - - ${OLLAMA_SSH_KEY_PUB:-/dev/null}:/root/.ollama/id_ed25519.pub:ro ports: - "${OLLAMA_PORT:-11434}:11434" shm_size: '${OLLAMA_SHM_SIZE:-2gb}' diff --git a/scripts/generate-olla-config.sh b/scripts/generate-olla-config.sh index 298dea4..e8fdaa7 100755 --- a/scripts/generate-olla-config.sh +++ b/scripts/generate-olla-config.sh @@ -43,7 +43,9 @@ if [[ -f "$ENV_FILE" ]]; then val="${val#\"}"; val="${val%\"}" val="${val#\'}"; val="${val%\'}" case "$key" in - OLLA_*|OLLAMA_REMOTE_*) eval "export $key=\$val" ;; + OLLA_*|OLLAMA_REMOTE_*) + declare -gx "$key=$val" + ;; esac done < "$ENV_FILE" fi