Skip to content

2026

How to Run Your Own Local AI with Ollama on Windows

Running AI models locally gives you full control over your data and eliminates dependency on cloud services. Ollama is a great service to serve and inference AI models locally, providing super fast speeds. In this guide, we'll set up Ollama with Open WebUI using Docker on Windows with WSL2.

Prerequisites

Before we begin, you'll need:

  • Windows 10/11 with WSL2 support
  • NVIDIA GPU (recommended for better performance)
  • Basic familiarity with command line

Hardware Requirements

When choosing which model to run, consider your hardware:

Model Size VRAM Required Recommended GPU
7B 8-16 GB NVIDIA RTX 3060
14B 24 GB NVIDIA RTX 3090/4090
32B 48+ GB Multi-GPU setups

Note

Models can also run on CPU with sufficient RAM (16-32 GB for 7B models), but performance will be significantly slower.