Create a cluster
FPT Cloud supports the following cards:
In the Hanoi and SGN regions, the A30 GPU card is supported.
In the Hanoi 2 and Japan regions, the following GPU cards are supported: H100 SXM5, H200 SXM5
Requirements
CPU, GPU, RAM, Storage, and Instance quotas: Must be sufficient for the desired Kubernetes cluster configuration. If using Autoscale, the number of GPUs must meet the desired maximum node count (note the Min node and Max node settings).
01 Network subnet: Network used for Kubernetes Nodes, the subnet must have a Static IP Pool.
Step-by-Step
GPU A30
Step 1: Access the FPT Cloud portal console.fptcloud.com, select Kubernetes, and click "Create a Kubernetes Engine".
Step 2: Enter the basic information for the cluster, then click the Next button:
Basic Information:
Name: Enter the cluster name.
Network: Subnet used to deploy Virtual Machines (VMs) of the Kubernetes Cluster.
Version: Select the version of the Kubernetes Cluster.
Cluster Endpoint Access: Option to access the Kubernetes cluster endpoint.
Public: The k8s cluster's Apiserver endpoint is accessible from the public. Connections from worker nodes to the apiserver go through the public path.
Public & Private: The k8s cluster's Apiserver endpoint is accessible from the public network. Connections from worker nodes to the apiserver go through the private network.
Private: The k8s cluster's Apiserver endpoint can only be accessed privately within the VPC. Connections from worker nodes to the apiserver go through the private network.
How to choose the appropriate Cluster Endpoint
Public: Use this option if customers need to access the Kubernetes API Endpoint from outside the VPC.
Public & Private: Use when you need a publicly accessible Endpoint and want to whitelist IPs accessing this apiserver endpoint.
Private: Use when only internal access to the Endpoint within the VPC is required.
Note:
Customers need to select the appropriate Cluster Endpoint Access based on the security requirements and network architecture of the system.
If Public & Private or Private is selected, an additional Allow CIDR field will appear to enter a list of
IP address ranges that have access to the Kubernetes Cluster Endpoint.
Explanation of Allow CIDR
CIDR (Classless Inter-Domain Routing) is a format for specifying the IP range that has access to the Kubernetes API Endpoint.
If left blank, the default is 0.0.0.0/0, meaning the Endpoint will be accessible from any IP address.
If you enter a specific value (e.g., 192.168.1.0/24), only IPs in the range 192.168.1.0 –
192.168.1.255 will be able to access.
Note: For high security, customers should restrict CIDR to allow only internal IP ranges instead of
0.0.0.0/0.
Step 3: Enter the information for the Kubernetes cluster to be created. In addition to the usual Kubernetes creation information, you need to select the configuration for the GPU in the Worker Group:
Select instance type: GPU
Select GPU type: NVIDIA Tesla A30
Select GPU sharing configuration
Select the GPU configuration type (CPU/RAM/GPU RAM)
Note:
In the "GPU Driver Installation Type" section, there are two options: Pre-installand User-install.
A driver is a program that allows the operating system to communicate with hardware, specifically in this case between the worker's operating system (Windows, Ubuntu, etc.) and the GPU. The operating system cannot use the GPU without a driver.
For the "Pre-install" option, the customer's cluster will have the Nvidia GPU driver automatically added.
For the "User-install" option, customers can manually install the GPU driver to choose the appropriate driver version.
Step 4: Click Create and review the initialization information.
Step 5: Monitor the Kubernetes cluster creation status. Once the status shows Successed (Running), proceed to use and deploy the application.
GPU H100 SXM5
Step 1: On the FPT Cloud Portal menu, select Containers> Kubernetes> Create a Kubernetes Engine.
Step 2: Enter the basic information for the cluster, then click the Nextbutton:
Basic Information:
Name: Enter the cluster name.
Network: Subnet used to deploy Kubernetes Cluster Virtual Machines (VMs).
Version: Select the version of the Kubernetes Cluster.
Cluster Endpoint Access: Select the Kubernetes cluster endpoint access option.
Step 3: Configure the Nodes Pool according to your needs, then click the Next button:
For the H100 card, the portal does not support creating GPU workers as the base worker group. Customers should create GPU workers starting from worker group 2 onwards.
Base worker group:
Instance Type: Select the General Instance type
Type: Select the configuration (CPU & Memory) for the Worker Nodes.
Container Runtime: Select Containerd.
Policy: Select the Storage Policytype (corresponding to IOPS) for the Worker Node Disk.
Disk: Select the root disk capacity for the Worker Nodes.
Scale min: Minimum number of Worker Node VM instances for the k8s cluster. The recommended minimum is 03 Nodes for the Production environment.
Scale max: The maximum number of Worker Node VM instances for a worker group in the k8s cluster.
Label: Apply a label to the Worker Group.
Worker Group n:
Select instance type: GPU
Select GPU type: NVIDIA H100 SXM5
Select GPU sharing configuration
Select GPU type configuration (CPU/RAM/GPU RAM)
Note:
In the "GPU Driver Installation Type" section, there are two options: Pre-installand User-install.
A driver is a program that allows the operating system to communicate with the hardware, specifically in this case between the worker's operating system (Windows, Ubuntu, etc.) and the GPU. The operating system cannot use the GPU without a driver.
For the "Pre-install" option, the customer's cluster will have the Nvidia GPU driver automatically added.
For the "User-install" option, customers can manually install the GPU driver to select the appropriate driver version.
Step 4: Click Create and review the initialization information.
Step 5: Monitor the Kubernetes cluster creation status. Once the status shows "Successed (Running)," proceed to use and deploy the application.
GPU H200 SXM5
Step 1: On the FPT Portal menu, select Containers> Kubernetes> Create a Kubernetes Engine.
Step 2: Enter the basic information for the cluster, then click the Next button:
Basic Information:
Name: Enter the Cluster name.
Network: Subnet used to deploy Kubernetes Cluster Virtual Machines (VMs).
Version: Select the version of the Kubernetes Cluster.
Cluster Endpoint Access: Option to access the Kubernetes cluster endpoint.
Note:
Customers need to select the appropriate Cluster Endpoint Access based on the security requirements and network architecture of the system.
If Public & Private or Private is selected, an additional Allow CIDR field will appear to enter a list of IP address ranges that have access to the Kubernetes Cluster Endpoint.
Step 3: Configure the Nodes Pool according to your usage needs, then click the Next button:
For the GPU H200, the portal does not support creating GPU workers as worker group bases. Customers are kindly requested to create GPU workers starting from worker group 2 onwards.
Worker Group base:
Instance Type: Select General Instance Type
Type: Select the configuration (CPU & Memory) for the Worker Nodes.
Container Runtime: Select Containerd.
Policy: Select the Storage Policytype (corresponding to IOPS) for the Worker Node Disk.
Disk: Select the root disk capacity for Worker Nodes.
Scale min: Minimum number of Worker Node VM instances for the k8s cluster. The recommended minimum is 03 Nodes for the Production environment.
Scale max: The maximum number of Worker Node VM instances for a worker group in the cluster.
Worker Group n:
Label: Assign a label to the Worker Group
Select instance type: GPU
Select GPU type: NVIDIA H200 SXM5
Select GPU sharing configuration
Select GPU configuration type (CPU/RAM/GPU RAM)
Note:
In the "GPU Driver Installation Type" section, there are two options: Pre-installand User-install.
A driver is a program that allows the operating system to communicate with hardware, specifically in this case between the worker's operating system (Windows, Ubuntu, etc.) and the GPU. The operating system cannot use the GPU without a driver.
For the "Pre-install" option, the customer's cluster will have the Nvidia GPU driver automatically added.
For the "User-install" option, customers can manually install the GPU driver to choose the appropriate driver version.
Step 4: Click Create and review the initialization information.
Step 5: Monitor the Kubernetes cluster creation status. Once the status shows Successed (Running), proceed to use and deploy the application.
Last updated
