Serving DeepSeek-R1

Using Ollama and Open WebUI

Ollama is an open-source tool that enables running, managing, and customizing large language models (LLMs) on personal computers or servers, supporting various models such as Llama, DeepSeek, Mistral, and more. Open-WebUI is an open-source web interface specifically designed to

Interact with Ollama, providing a user-friendly experience for managing and using LLM models.

This document will guide you through the steps to deploy the DeepSeek-R1 model on the FPT Managed GPU Cluster using Ollama and Open-WebUI so that users can use it simply and easily.

Step 1: Clone the existing Open-WebUI source code and script

git clone https://github.com/open-webui/open-webui
cd open-webui/kubernetes

Step 2: Run the scripts to deploy ollama and open-webui. The directory contains all the necessary files for deployment, such as namespace, ollama statefulSet, ollama service, open-webui deploymentand open-webui service.

cd kubernetes
kubectl apply -f ./kubernetes/manifest

Step 3: Access open-webui on the browser at the forwarded port, for example: http://localhost:52433. For the first time installing and using OpenWebUI, users will need to configure the following information: name, email, password.

Step 4: After installation is complete, users select the model to use. For example, here we will install the DeepSeek-R1 model, version 1.5b.

Step 5: Once the model has been loaded and run, users can interact with the model very simply and intuitively through the interface.

Last updated