What You Need

🖥️
Windows PC

Windows 10 or 11

🎮
NVIDIA GPU

8GB+ VRAM minimum

🐳
Docker Desktop

Free download required

⏱️
~10 Minutes

Setup time

Pick the Right Model for Your GPU

The model you run depends on your GPU's VRAM. Here's exactly what to use:

GPU VRAM Recommended Model Best For
4 GB qwen2.5:1.5b Basic tasks, simple Q&A
8 GB ⭐ qwen3-vl:8b Daily assistant + image analysis
12 GB qwen2.5:14b Smarter responses, better coding
16 GB deepseek-r1:14b Complex reasoning tasks
24 GB qwen2.5-coder:32b GPT-4 level coding
32 GB+ qwen2.5:72b Frontier model quality

💡 Most people start with 8GB and qwen3-vl:8b. It's fast, capable, and handles images too.

1 Download & Install Ollama

Ollama is the engine that runs AI models locally. It handles everything under the hood.

  1. Go to ollama.com/download
  2. Click Download for Windows
  3. Run the installer — click through and let it install
  4. Ollama will open automatically when done
📝 Note: Ollama runs as a background service. You'll see it in your system tray after install. This is what serves your AI models locally.

2 Install Docker Desktop

Docker runs the Open WebUI interface — the beautiful chat UI you'll actually talk to.

  1. Go to docker.com/products/docker-desktop
  2. Download and install Docker Desktop for Windows
  3. Restart your PC if prompted
  4. Open Docker Desktop and make sure it's running (whale icon in system tray)

3 Launch Open WebUI

Open PowerShell as Administrator and run this single command:

PowerShell (Run as Administrator)
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
📝 Note: This command downloads and starts Open WebUI. It only needs to run once — Docker will auto-start it on reboot.

Then open your browser and go to:

4 Create Your Account & Pick a Model

  1. Click Sign Up to create a local admin account (stays on your machine)
  2. Once logged in, click the model dropdown at the top
  3. Search for your model (e.g. qwen3-vl:8b) and click to download it
  4. Wait for the download to complete — then start chatting!
🎉 That's it. You're now running a free, private AI on your own hardware. No subscriptions. No data leaving your machine.

5 Enable Persistent Memory

Make your AI remember things between conversations — just like Claude or ChatGPT with memory enabled.

  1. Click your profile icon (top right)
  2. Go to Settings → Personalization
  3. Toggle on Memory (Experimental)

Now you can tell it things like "commit this to your memory" or "remember my preferences" and it will persist across sessions.

All Commands & Links

Questions? Join the Community

Drop a comment on the video or come chat in the AlphaMine Discord. We're happy to help you get your local AI setup dialed in.