This repository contains Docker configuration for running Ollama with Open WebUI, providing a user-friendly interface for interacting with various AI models.
- Docker and Docker Compose installed on your system
- NVIDIA GPU with updated NVIDIA drivers
- NVIDIA Container Toolkit installed (Installation Guide)
- Clone this repository
- Start the Docker containers:
docker compose up -dThis will:
- Start Ollama server with GPU support
- Launch Open WebUI interface
- Create persistent volumes for both services
Once the containers are running, you can access the Open WebUI interface at: http://localhost:3000
To pull the Deepseek model, use either:
- Via CLI:
docker exec -it ollama ollama pull deepseek-coder- Via Open WebUI:
- Navigate to http://localhost:3000
- Click on "Models" in the sidebar
- Search for "deepseek-coder"
- Click "Pull" button
To pull the Qwen model, use either:
- Via CLI:
docker exec -it ollama ollama pull qwen2.5- Via Open WebUI:
- Navigate to http://localhost:3000
- Click on "Models" in the sidebar
- Search for "qwen"
- Select version 2.5
- Click "Pull" button
- Minimum 8GB RAM (allocated to Docker)
- 4 CPU cores
- NVIDIA GPU with at least 8GB VRAM
- Sufficient storage space for models:
- Deepseek-coder: ~4GB
- Qwen 2.5: ~4GB
If you encounter issues:
- Check Docker logs:
docker compose logs -f- Ensure NVIDIA drivers are properly installed:
nvidia-smi- Verify GPU access in container:
docker exec -it ollama nvidia-smiThe setup includes several security measures:
- Container memory limited to 8GB
- CPU usage limited to 4 cores
- Read-only filesystem
- Dropped capabilities and no privilege escalation
- Automatic restart on failure