Want to deploy powerful open-source AI models like Qwen 2.5, Llama 3, and DeepSeek-R1 locally, but struggling to find a simple and easy-to-use method?
Don't worry! The golden combination of Ollama + Open WebUI will clear all obstacles for you.
This article will provide a step-by-step tutorial, detailing how to easily build a local AI environment using Ollama + Open WebUI, giving you an exclusive, powerful AI assistant to explore the endless possibilities of AI!
Warm Reminder: Due to hardware limitations, local deployment usually cannot run the largest version of DeepSeek-R1 (such as 67B). But don't worry, smaller-scale models (such as 1.3B or 7B) can run smoothly on most personal computers and provide excellent reasoning capabilities. More importantly, you can choose the version that best suits your needs!
Why Choose Ollama + Open WebUI?
Among many local deployment solutions, the Ollama + Open WebUI combination stands out as the preferred choice for many AI enthusiasts. What is their charm?
- Ollama: A Simplified Model Engine
- Ollama is like an "AI model treasure chest". With just one command, you can download, install, and run various mainstream large language models, such as Llama 3 and DeepSeek-R1!
- Open WebUI: An Elegant and Easy-to-Use Interactive Interface
- Open WebUI adds a gorgeous layer to Ollama. It provides a beautiful and intuitive web interface.
- Completely open source and free.
After deployment, just open http://127.0.0.1:8080
in your browser to start a conversation with your AI assistant:
Exclusive for Windows Users: One-Click Startup Integration Package, Say Goodbye to Cumbersome Configurations!
Considering the difficulties Windows users may encounter when configuring the Docker environment, we have thoughtfully prepared an integration package that can be used by downloading and decompressing, truly achieving "out-of-the-box" functionality!
Download and Decompress the Integration Package:
Integration Package Download Address: https://www.123684.com/s/03Sxjv-4cTJ3
- If you have not installed Ollama, please double-click the
ollama-0.1.28-setup.exe
file in the integration package to install it. The installation process is very simple, just click "Next" all the way.
- If you have not installed Ollama, please double-click the
Start WebUI:
- Double-click the
启动webui.bat
file in the integration package to start Open WebUI.
- The first time you start it, the system will prompt you to set up an administrator account. Please follow the prompts to complete the registration.
- Double-click the
Select the Model You Want to Use
After entering Open WebUI, you will see the model selection area in the upper left corner. If there is no model in the list, don't worry, it means you haven't downloaded any models yet.
You can directly enter the model name in the input box and download it online from Ollama.com:
Model Selection Tips:
- Model Treasury: Visit https://ollama.com/models to browse the rich model resources officially provided by Ollama.
- Parameter Scale: Each model has different versions (e.g., 1.3B, 7B, 67B, etc.), representing different parameter scales. The more parameters, the more powerful the model usually is, but it also requires more computing resources (memory and video memory).
- Do What You Can: Choose the right model based on your hardware configuration. In general, if your "memory + video memory" size is greater than the model file size, you can run the model smoothly.
- Deepseek-R1 Selection: Search for
deepseek-r1
in Ollama's model library to find it.
Taking the deployment of the deepseek-r1
model as an example:
Select the Model Specification: On the https://ollama.com/library page, find the model version you want to deploy (e.g.,
deepseek-r1
).Download the Model: Paste the model name (e.g.,
deepseek-r1
) into the input box in the upper left corner of Open WebUI, click the "Pull from ollama.com" button to start the download.Wait for the Download to Complete: The download time depends on your network speed and the size of the model, please be patient.
Start Your AI Journey
Once the model is downloaded, you can chat freely with DeepSeek-R1 in Open WebUI! Explore its powerful features!
If the model supports it, you can also upload images, files, etc. for multimodal interaction. Let your AI assistant not only speak but also "read pictures"!
Advanced Exploration: Open WebUI's Hidden Treasures
Open WebUI has more than just these features! Click the menu button in the upper left corner and you'll find more surprises:
Personalization: In the "Settings" panel, you can customize the interface theme, font size, language, etc. to create your own AI interaction experience.
- You can also customize prompts to make the AI assistant understand you better!
Multi-User Management: In the "Admin" panel, you can set user registration methods, permissions, etc. to facilitate multiple people sharing your local AI resources.
Adjust Detailed Parameters: Click in the upper right corner to set advanced parameters.
Multi-Model Comparison: Which One is Better?
Open WebUI also supports multi-model comparison, allowing you to easily compare the output results of different models and find the one that best suits your needs!
GPU Acceleration: Squeeze Out Your Graphics Card Performance! (Optional)
If you have an NVIDIA graphics card and have installed the CUDA environment, then congratulations, you can use simple operations to let Ollama use GPU to accelerate model reasoning and greatly improve the response speed of the AI assistant!
- Double-click the
GPU-cuda支持.bat
file in the integration package to install CUDA dependencies.
Ollama + Open WebUI, this golden combination, opens a door to the local AI world for you. Now, you can get rid of cloud constraints, create your own AI think tank, and explore the unlimited possibilities of AI!