User Tools

Site Tools


ai:openwebui:start

Setting up OpenWebUI and Ollama using Docker

Prerequisites

1. A system with Docker installed

Step 1: Install Ollama

1. Open a terminal window.

2. Run the following command to install Ollama:

 curl https://ollama.ai/install.sh | sh

3. Verify the installation by running:

 ollama --version

Step 2: Pull a Language Model

Pull a language model (e.g., Llama 3) using Ollama:

 ollama pull llama3

Step 4: Set Up OpenWebUI

1. Open a new terminal window.

2. Run the OpenWebUI Docker container:

 sudo docker run -d \
 --network=host -v /home/azam/open-webui/:/app/backend/data \
 -e OLLAMA_BASE_URL=http://127.0.0.1:11434 \
 --name open-webui --restart always ghcr.io/open-webui/open-webui:main

This command does the following:

1. Runs the container in detached mode (`-d`)

2. Names the container “openwebui” (`–name openwebui`)

3. If Ollama and OpenWebUI are running on the same system it uses the system network (`–network=host`)

4. Sets the Ollama API base URL environment variable (`-e OLLAMA_API_BASE_URL=…`)

6. Uses the OpenWebUI Docker image (`ghcr.io/open-webui/open-webui:main`)

Step 5: Access OpenWebUI

1. Open a web browser and navigate to:

 http://localhost:8080

Step 6: Configure OpenWebUI

1. In the OpenWebUI interface, first create an account.

2. Login to the account created.

Step 7: Test the Setup

1. Start a new chat in OpenWebUI.

2. Select the language model you pulled earlier (e.g., Llama 3).

3. Send a test message to verify that everything is working correctly.

Additional Docker Commands

- To stop the OpenWebUI container:

docker stop openwebui

- To start the OpenWebUI container again:

docker start openwebui

- To remove the OpenWebUI container:

docker rm openwebui
ai/openwebui/start.txt · Last modified: 2024/11/25 12:26 by azam01