Set up a local AI like ChatGPT on your computer (Web Interface)
Vladislav G.
Marketing Director & Lead Web Developer | Driving Data-Driven Marketing Strategies with Technical Expertise
Did you know you can run powerful AI models right on your computer? It’s true! Today, I will show you how easy it is to get started with Llama 3.2 and OpenWebUI.
Watch on YouTube: Run AI On YOUR Computer
Running Llama 3.2 Locally: A Comprehensive Guide
Introduction to Llama 3.2
Llama 3.2 is the latest iteration of Meta’s open-source language model, offering enhanced capabilities for text and image processing. It is designed to run efficiently on local devices, making it ideal for applications that require privacy and low latency. The model comes in various sizes, including 1B, 3B, and 11B parameters. In this tutorial, I’m going to use the 1B model, but you can download any you like.
Before setting up Llama 3.2 locally, ensure you have the following:
Step-by-Step Installation Guide
Step 1: Install Docker
Docker is a tool that allows developers to package applications and their dependencies into a standardized unit called a container, which can run consistently across different computing environments. Unlike virtual machines, containers are lightweight and share the host system’s operating system, making them more efficient and faster to start.
You can download the Docker Desktop application from the Docker website https://www.docker.com/ or via the terminal. Docker is available for Mac, Windows, and Linux.
How to Install Docker on Ubuntu
Download the package https://desktop.docker.com/linux/main/amd64/docker-desktop-amd64.deb
Then use these commands:
sudo apt-get update
sudo apt-get install [path to the docker package].deb
More details: https://docs.docker.com/desktop/install/linux/ubuntu/
Step 2: Install Ollama
Ollama is essential for running large language models like Llama 3.2 locally.
Follow these steps to install it:
I’m going to use the terminal and this command:
curl -fsSL https://ollama.com/install.sh | sh
Step 3: Download a New Llama 3.2 model from Ollama website
Once Ollama is set up, you can download Llama 3.2 models. All the available models you can find here https://ollama.com/library
领英推荐
Since I want to use 1B parameters model I’ll be using this command:
ollama run llama3.2:1b
Step 4: Install OpenWebUI with Default Configuration
OpenWebUI is a self-hosted, extensible web interface designed to interact entirely offline with large language models (LLMs). It offers a user-friendly experience similar to ChatGPT, supports integration with various LLMs, such as those compatible with OpenAI and Ollama, and provides features like markdown support, model management, and multi-user access.
Go to https://docs.openwebui.com/getting-started/ and find the section Quick Start with Docker. Copy the code and then run it in the terminal.
I’m going to run this code:
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama
We will run a Docker command. It is used to run a containerized version of the OpenWebUI application.
Once the installation is finished, open your browser and go to https://0.0.0.0:3000/. If everything is okay, you will see the website. Create an account, and then you will see the welcome screen, which is similar to ChatGPT.
Troubleshooting Common Issue
WebUI models not being pulled/applied
Some of you could face an issue when the models are not available in the drop-down. To fix it, follow these instructions:
2. Click on “Admin Settings”.
3. Click on “Models.”
4. In the input field, write the name of the model you would like to pull and click on the “download” icon.
All the names you can find on https://ollama.com/library/llama3.2
This should fix the issue.
Conclusion
Running Llama 3.2 locally provides significant advantages regarding privacy and control over AI applications. But to have a smooth experience, you would need a powerful computer. :)
Thanks! :)