Serving DeepSeek-R1
Using Ollama and Open WebUI
Ollama is an open-source tool that enables running, managing, and customizing large language models (LLMs) on personal computers or servers, supporting various models such as Llama, DeepSeek, Mistral, and more. Open-WebUI is an open-source web interface specifically designed to
Interact with Ollama, providing a user-friendly experience for managing and using LLM models.
This document will guide you through the steps to deploy the DeepSeek-R1 model on the FPT Managed GPU Cluster using Ollama and Open-WebUI so that users can use it simply and easily.
Step 1: Clone the existing Open-WebUI source code and script
git clone https://github.com/open-webui/open-webui
cd open-webui/kubernetesStep 2: Run the scripts to deploy ollama and open-webui. The directory contains all the necessary files for deployment, such as namespace, ollama statefulSet, ollama service, open-webui deploymentand open-webui service.
cd kubernetes
kubectl apply -f ./kubernetes/manifestStep 3: Access open-webui on the browser at the forwarded port, for example: http://localhost:52433. For the first time installing and using OpenWebUI, users will need to configure the following information: name, email, password.
Step 4: After installation is complete, users select the model to use. For example, here we will install the DeepSeek-R1 model, version 1.5b.
Step 5: Once the model has been loaded and run, users can interact with the model very simply and intuitively through the interface.
Last updated
