How to Install and Use Ollama WebUI on Windows

Ollama WebUI is a revolutionary LLM local deployment framework with chatGPT like web interface. Let’s get chatGPT like web ui interface for your ollama deployed LLMs.

What is Open WebUI (Formerly Ollama WebUI)?

Open WebUI is an extensible, feature-rich, and user-friendly self-hosted WebUI designed to operate entirely offline. It supports various LLM runners, including Ollama and OpenAI-compatible APIs.

Ollama WebUI is a revolutionary LLM local deployment framework with chatGPT like web interface. Let’s get chatGPT like web ui interface for your ollama deployed LLMs. Just follow these 5 steps to get up and get going.

System Requirements

Windows 10 64-bit: Minimum required is Home or Pro 21H2 (build 19044) or higher, or Enterprise or Education 21H2 (build 19044) or higher.

Windows 11 64-bit: Home or Pro version 21H2 or higher, or Enterprise or Education version 21H2 or higher.

WSL version 1.1.3.0 or later. Turn on the WSL 2 feature on Windows.

8 cores 64-bit processor, 16GB+ RAM, Nvidia Graphics Card with 4GB+ VRAM

Docker Desktop, the latest version

5 Steps to Install and Use Ollama Web UI

Digging deeper into Ollama and Ollama WebUI on a Windows computer is an exciting journey into the world of artificial intelligence and machine learning. This detailed guide walks you through each step and provides examples to ensure a smooth launch.

Step 1 - Install Ollama
install ollama on windows

Download Ollama from https://ollama.com/download/windows, then right click on the downloaded OllamaSetup.exe file and run the installer as administrator. Once the installation is complete, Ollama is ready to use on your Windows system. An Ollama icon will be added to the tray area at the bottom of the desktop.

To run Ollama and start utilizing its AI models, you'll need to use a terminal on Windows. We'll skip it here and let's see how to install WebUI for a better experience.

Step 2 - Install Ollama WebUI

Run the below docker command to deploy ollama-webui docker container on your local machine. If Ollama is on your computer, use this command:

docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main

To connect to Ollama on another server, change the OLLAMA_BASE_URL to the server's URL. So if Ollama is on a Different Server, use this command:

docker run -d -p 3000:8080 -e OLLAMA_BASE_URL=https://example.com -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
install ollama webui

Note: When using Docker to install Open WebUI, make sure to include the -v open-webui:/app/backend/data in your Docker command. This step is crucial as it ensures your database is properly mounted and prevents any loss of data.

Step 3 - Sign in to Open WebUI

After installation, you can access Open WebUI at http://localhost:3000. For the first time you need to register by clicking "Sign up".

sign in to open webui

Once registered, you will be routed to the home page of ollama-webui.

open webui home page
Step 4 - Pull a model from Ollama.com

Click the settings icon in the upper right corner to open the settings window, as shown in the figure below, enter the model tag(e.g. llama2:7b, gemma:2b), click the download button on the right, and wait for the model to be downloaded.

open ui manage ollama models
Step 5 - Select a model and Enjoy your AI chat

As shown below, we have downloaded the gemma:2b and llama2:7b models, select gemma:2b.

Select Gemma:2b model

Enter the prompt word "What is the future of AI?" and press Enter to send a message.

chat with Gemma 2b model

Conclusion

In this tutorial, we cover the basics of getting started with Ollama WebUI on Windows. Ollama stands out for its ease of use, automatic hardware acceleration, and access to a comprehensive model library. Ollama WebUI is what makes it a valuable tool for anyone interested in artificial intelligence and machine learning.

Additional - Some Good GPU Plans for Ollama AI

Professional GPU VPS - A4000

129.00/mo
1mo3mo12mo24mo
Order Now
  • 32GB RAM
  • 24 CPU Cores
  • 320GB SSD
  • 300Mbps Unmetered Bandwidth
  • Once per 2 Weeks Backup
  • OS: Linux / Windows 10
  • Dedicated GPU: Quadro RTX A4000
  • CUDA Cores: 6,144
  • Tensor Cores: 192
  • GPU Memory: 16GB GDDR6
  • FP32 Performance: 19.2 TFLOPS
  • Available for Rendering, AI/Deep Learning, Data Science, CAD/CGI/DCC.
Summer Sale

Advanced GPU - A4000

167.2/mo
Save 40% (Was $279.00)
1mo3mo12mo24mo
Order Now
  • 128GB RAM
  • Dual 12-Core E5-2697v2
  • 240GB SSD + 2TB SSD
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: Nvidia Quadro RTX A4000
  • Microarchitecture: Ampere
  • Max GPUs: 2
  • CUDA Cores: 6144
  • Tensor Cores: 192
  • GPU Memory: 16GB GDDR6
  • FP32 Performance: 19.2 TFLOPS
  • Good Choice for Android Emulators, 3D Rendering, Video Editing, AI/Deep Learning, Data Science, etc

Advanced GPU - A5000

269.00/mo
1mo3mo12mo24mo
Order Now
  • 128GB RAM
  • Dual 12-Core E5-2697v2
  • 240GB SSD + 2TB SSD
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: Nvidia Quadro RTX A5000
  • Microarchitecture: Ampere
  • Max GPUs: 2
  • CUDA Cores: 8192
  • Tensor Cores: 256
  • GPU Memory: 24GB GDDR6
  • FP32 Performance: 27.8 TFLOPS
  • High-Powered for Android Emulators, 3D Rendering, Video Editing, AI/Deep Learning, etc.

Enterprise GPU - RTX A6000

409.00/mo
1mo3mo12mo24mo
Order Now
  • 256GB RAM
  • Dual 18-Core E5-2697v4
  • 240GB SSD + 2TB NVMe + 8TB SATA
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: Nvidia Quadro RTX A6000
  • Microarchitecture: Ampere
  • Max GPUs: 1
  • CUDA Cores: 10,752
  • Tensor Cores: 336
  • GPU Memory: 48GB GDDR6
  • FP32 Performance: 38.71 TFLOPS

Enterprise GPU - A40

439.00/mo
1mo3mo12mo24mo
Order Now
  • 256GB RAM
  • Dual 18-Core E5-2697v4
  • 240GB SSD + 2TB NVMe + 8TB SATA
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: Nvidia A40
  • Microarchitecture: Ampere
  • Max GPUs: 1
  • CUDA Cores: 10,752
  • Tensor Cores: 336
  • GPU Memory: 48GB GDDR6
  • FP32 Performance: 37.48 TFLOPS
  • Higher-Powered for 3D Rendering, Video Editing, AI/Deep Learning, etc.
New Arrival

Multi-GPU - 3xRTX A5000

539.00/mo
1mo3mo12mo24mo
Order Now
  • 256GB RAM
  • Dual 18-Core E5-2697v4
  • 240GB SSD + 2TB NVMe + 8TB SATA
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: 3 x Quadro RTX A5000
  • Microarchitecture: Ampere
  • Max GPUs: 3
  • CUDA Cores: 8192
  • Tensor Cores: 256
  • GPU Memory: 24GB GDDR6
  • FP32 Performance: 27.8 TFLOPS
New Arrival

Multi-GPU - 3xRTX A6000

899.00/mo
1mo3mo12mo24mo
Order Now
  • 256GB RAM
  • Dual 18-Core E5-2697v4
  • 240GB SSD + 2TB NVMe + 8TB SATA
  • 100Mbps-1Gbps
  • OS: Windows / Linux
  • GPU: 3 x Quadro RTX A6000
  • Microarchitecture: Ampere
  • Max GPUs: 3
  • CUDA Cores: 10,752
  • Tensor Cores: 336
  • GPU Memory: 48GB GDDR6
  • FP32 Performance: 38.71 TFLOPS
Let us get back to you

If you can't find a suitable GPU Plan, or have a need to customize a GPU server, or have ideas for cooperation, please leave me a message. We will reach you back within 36 hours.

Email *
Name
Company
Message *
I agree to be contacted as per Database Mart privacy policy.