diff --git a/Ollama/docker-compose.yml b/Ollama/docker-compose.yml new file mode 100644 index 0000000..b503635 --- /dev/null +++ b/Ollama/docker-compose.yml @@ -0,0 +1,44 @@ +version: '3.6' + +services: + ollama: + # Uncomment below for GPU support + # deploy: + # resources: + # reservations: + # devices: + # - driver: nvidia + # count: 1 + # capabilities: + # - gpu + volumes: + - ollama:/root/.ollama + # Uncomment below to expose Ollama API outside the container stack + # ports: + # - 11434:11434 + container_name: ollama + pull_policy: always + tty: true + restart: unless-stopped + image: ollama/ollama:latest + + ollama-webui: + build: + context: . + args: + OLLAMA_API_BASE_URL: '/ollama/api' + dockerfile: Dockerfile + image: ollama-webui:latest + container_name: ollama-webui + depends_on: + - ollama + ports: + - 3000:8080 + environment: + - "OLLAMA_API_BASE_URL=http://ollama:11434/api" + extra_hosts: + - host.docker.internal:host-gateway + restart: unless-stopped + +volumes: + ollama: {} diff --git a/Ollama/readme.md b/Ollama/readme.md new file mode 100644 index 0000000..7621e5d --- /dev/null +++ b/Ollama/readme.md @@ -0,0 +1,5 @@ +1. Clone the repo from: https://github.com/ollama-webui/ollama-webui +2. Tweak the docker-compose to your liking +3. Run the container: sudo docker compose up -d + +Let it build :) \ No newline at end of file